Moss 2007: Full Crawl TraceLog shows errors

When I   full crawled my MOSS 2007, TraceLog says "....... value is not   available" in every place many many times.
I found the error log at   TopSite, pages, subsites, Lists, Document and FileSharedLibrary(by Doc Ave)
The error is like that.
DocFormat value is not   available for url:...................
CLSID value is not available   for url: ...................
RedirectedURL value is not   available for url:...................
BindToStream is not available   for url:...................
FileName value is not   available for url:...................
CSTS3Accessor::GetItemSecurityDescriptor:   Return error to caller, hr=80041211...................
There are no more   chunks,...................
Does anyone know what these   errors mean and they are critical error or not?
I tried to search these errors but I couldn't find any information would help me.

I think you dont have mysite as sts3 stands for mysite
Here is a exact issue
http://webcache.googleusercontent.com/search?q=cache:VMcltdpjfm0J:share-point.blogspot.com/2008/11/moss2007-crawl-issue-after-installation.html+&cd=2&hl=en&ct=clnk&gl=in
Perform the steps mentioned in the KB article below to fix this issue:
http://support.microsoft.com/kb/896861
If this helped you resolve your issue, please mark it Answered. You can reach me through http://freeit-support.com/

Similar Messages

  • While crawling content from Moss 2007 to Sp2013 its throwing error

    hi,
    When i am trying to crawl the Moss 2007 site content in SP2013 serach I am getting below Error in the Crawl error log.
    I have provided owner access for the Moss 2007 content DB,to the user accout which i am using in SP2013 serach.And also provided read access for this account to the Moss 2007.But still i am unable to crawl.
    1 The URL of the item could not be resolved. The repository might be unavailable, or the crawler proxy settings are not configured. To configure the crawler proxy settings, use Search Administration page.
    2 This item could not be crawled because the repository did not respond within the specified timeout period. Try to crawl the repository at a later time, or increase the timeout value on the Proxy and Timeout page in search administration. You might
    also want to crawl this repository during off-peak usage times.
    3 The crawler could not communicate with the server. Check that the server is available and that the firewall access is configured correctly. If the repository was temporarily unavailable, an incremental crawl will fix this error.
    4 Access is denied. Verify that either the Default Content Access Account has access to this repository, or add a crawl rule to crawl this repository. If the repository being crawled is a SharePoint repository, verify that the account you are using
    has "Full Read" permissions on the SharePoint Web Application being crawled.
    Thanks & Regards, Krishna

    Try to use website as content source to crawl it
    Create disableloopbackcheck registry on all sharepoint servers
    Are you able to browse sharepoint sites from crawl server. 
    check 
    http://blogs.technet.com/b/josebda/archive/2007/03/19/crawling-sharepoint-sites-using-the-sps3-protocol-handler.aspx
    http://sharepoint.stackexchange.com/questions/93691/using-sharepoint-2013-to-crawl-sharepoint-2007
    If you are still struggling you can watch the requests by using Fiddler as a proxy for the crawl process. Then you can kick off the crawl and watch the requests and the response from the 2007 server.

  • MOSS 2007 Search - Crawling is taking longer than usual time since last month for same content sources

    Hi all,
    Off late we have discovered that content crawling is taking longer than expected also overlapping to next scheduled too. Literally no crawl logs seen for hours. No entry in crawl logs. Is there anyone out here having similar issue? Please share a solution
    if any found.
    My farm is implemented with MOSS 2007 SP2 Ver no 12.0.0.6554
    There is not packet drop between index server, App and SQL server/Cluster
    Thank you in advance,
    Reach Ram
    Ramakrishna Pulipati SharePoint Consultant, Bangalore, INDIA

    I believe this is ready for submission for the Time Machine forum.
    As noted, it does not cover diagnosis and correction of specific problems or errors, as that would seem to be better handled separately.
    It also doesn't cover anything about Time Capsule, for the very good reason that I'm not familiar with them. If someone wants to draft a separate post for it, or items to add/update here, that would be great!
    Thanks very much.

  • "Unable to connect to the remote server" error message (ID 6102) in event view after installing MOSS 2007

     Hi brothers,
    After installed the Moss 2007(basic) , I find some error in event view. The event ID is "6102". More detail information is shown below.
     Event Type:        Error
    Event Source:    Office SharePoint Server
    Event Category:                Launcher Service Event ID:              6102Date:                     11/27/2007Time:                     8:22:17 AMUser:                     N/AComputer:          MOSS-001Description:LoadBalancer.RegisterLauncher failed:  Unable to connect to the remote server For more information, see Help and Support Center at http://go.microsoft.com/fwlink/events.asp. 
    Please help me, thanks.

    Hi,
    i think this has to do with the "Document Conversions Load Balancer Service".
    If you don't need the service than try to stop it:
    Start Central Administration > Operations > Services and stop the service.
    or 
    If you need it the service, try this:
    Start Central Administration > Operations > Service Account
    Change the service account to "Document Conversions Load Balancer Service" And Select a "Network Services" Click Save
    Hope this helps!
    Cheers,
    Daniel Bugday
    Web:
    SharePoint Forum
    Blog:
    Daniel Bugday's SharePoint Blog

  • When I full crawl my MOSS 2007 site, it is Intializing my site repeatedly!

    Hi everyone
    I'm using Moss 2007. When I did full crawl, it put out lots of logs. I read one of the logs and found that several strange parts on them.
    It sounds like ...
    1. It is initializing security for my subsite repeatedly.
    2. Cache is full
    3. My FileSharedLibrary allows no one to read items.
    1. It is initializing security for my subsite repeatedly.
    I don't know why it is initializing my subsite again and again. I believe this made the logs so large...
    Here is the log.
    Init security for site "MyItem'sURL" - File:d:\office\source\search\search\gather\protocols\sts3\sts3util.cxx Line:842
    Init Group GROUPNAME123 Id 7 - File:d:\office\source\search\search\gather\protocols\sts3\sts3util.cxx Line:3086 
    New User USERNAME123 Id 1 - File:d:\office\source\search\search\gather\protocols\sts3\sts3util.cxx Line:1073
    New User USERNAME456 Id 13 - File:d:\office\source\search\search\gather\protocols\sts3\sts3util.cxx Line:1073 
    2. Cache is full
    Also I think the log says the cache is full. But I have no idea this cache problem and the initializing problem are effecting each other or not.
    Remove V3 Cache, reason = oversize, committed 649506816 - File:d:\office\source\search\search\gather\protocols\sts3\sts3util.cxx Line:127 
    Remove V3 cache - File:d:\office\source\search\search\gather\protocols\sts3\sts3util.cxx Line:76 
    V3 Cache statistics: committed=649506816, UserCatalogCaches=1 - File:d:\office\source\search\search\gather\protocols\sts3\sts3util.hxx Line:172 
    3. My FileSharedLibrary access control allows no one to read items.
    Thirdly, I got this error many times. (FileSharedLibrary is the feature of DocAve)
    I searched "hr=80041211" and found MicroSoft said "Item will not be indexed. Its access control list allows no one to read the item. Used in GetSecurityDescriptor method of the IUrlAccessor interface.
    CSTS3Accessor::GetItemSecurityDescriptor: Return error to caller, hr=80041211
    - File:d:\office\source\search\search\gather\protocols\sts3\sts3acc.cxx Line:1210 
    I already checked DisabledLoopBackCheck is disabled, Crawl Account have a read access of my site properly, and I'm sure we have enough free space at C and D drives.
    Actually I can use my SharePoint site normally with no problem and it does not show any error on my site. But I really need to know the reason why this three points are happening...in order to avoid a huge-size log gives damage to server when I full
    crawl next time...
    Hope someone could help me to solve this problems

    Thank you Jaggi
    I think this is the almost same page that i checked already and mensioned in my question
    "hr=80041211" and found MicroSoft said "Item will not be indexed. Its access control list allows no one to read the item. Used in GetSecurityDescriptor method of the IUrlAccessor interface.
    Do you think my problem is really because of anything Access control problem..?

  • Sharepoint 2013 discussion board item view (flat.aspx) throws js error after being migrated from moss 2007

    After preforming a migration from moss 2007 to 2013 on-prem we have noticed that html tags are showing up in the discussion board posts.  I have tracked the problem down to the flat.aspx page throwing a js error. 
    'SPAnimationUtility' is undefined - coming from the sp.ui.discussions.js file
    this is what I see on the page: 
    The reply and edit buttons function fine.  Its just that the html is exposed. 
    Anyone seen this before or have a clue as to how to solve?  
    Thanks,
    -Alex

    As I remember there are a couple managed properties (metadata) that are used by the views that don't get created when you migrate a discussion board from a previous version.  I don't remember the specific properties, but take a look at a new discussion
    board in 2013 and compare the properties in the views and you will find them.  
    Paul Stork SharePoint Server MVP
    Principal Architect: Blue Chip Consulting Group
    Blog: http://dontpapanic.com/blog
    Twitter: Follow @pstork
    Please remember to mark your question as "answered" if this solves your problem.

  • Http service too busy error for SSIS adapters for SharePoint - SSIS 2005 and MOSS 2007

    Hi,
    a my old SSIS 2005 pkg that uses the SSIS adapters for SharePoint from only some days ago goes in error with this message for a data flow task that reads from a MOSS 2007 list:
    Error: System.ServiceModel.ServerTooBusyException: The HTTP service located at
    http://mysitename/_vti_bin/lists.asmx is too busy.  ---> System.Net.WebException: The remote server returned an error: (503) Server Unavailable.    at System.Net.HttpWebRequest.GetResponse()   
    at System.ServiceModel.Channels.HttpChannelFactory.HttpRequestChannel.HttpChannelRequest.WaitForReply(TimeSpan timeout)    --- End of inner exception stack trace ---  Server stack trace:     at System.ServiceModel.Channels.HttpChannelUtilities.ProcessGetResponseWebException(WebException
    webException, HttpWebRequest request, HttpAbortReason abortReason)    at System.ServiceModel.Channels.HttpChannelFactory.HttpRequestChannel.HttpChannelRequest.WaitForReply(TimeSpan timeout)    at System.ServiceModel.Channels.RequestChannel.Request(Message
    message, TimeSpan timeout)    at System.ServiceModel.Dispatcher.RequestChannelBinder.Request(Message message, TimeSpan timeout)    at System.ServiceModel.Channels.ServiceChannel.Call(String action, Boolean oneway, ProxyOperationRuntime
    operation, Object[] ins, Object[] outs, TimeSpan timeout)    at System.ServiceModel.Channels.ServiceChannelProxy.InvokeService(IMethodCallMessage methodCall, ProxyOperationRuntime operation)    at System.ServiceModel.Channels.ServiceChannelProxy.Invoke(IMessage
    message)  Exception rethrown at [0]:     at System.Runtime.Remoting.Proxies.RealProxy.HandleReturnMessage(IMessage reqMsg, IMessage retMsg)    at System.Runtime.Remoting.Proxies.RealProxy.PrivateInvoke(MessageData&
    msgData, Int32 type)    at Microsoft.Samples.SqlServer.SSIS.SharePointUtility.ListsService.ListsSoap.GetListAndView(GetListAndViewRequest request)    at Microsoft.Samples.SqlServer.SSIS.SharePointUtility.ListsService.ListsSoapClient.ListsService_ListsSoap_GetListAndView(GetListAndViewRequest
    request)    at Microsoft.Samples.SqlServer.SSIS.SharePointUtility.ListsService.ListsSoapClient.GetListAndView(String listName, String viewName)    at Microsoft.Samples.SqlServer.SSIS.SharePointUtility.Adapter.ListsAdapter.GetSharePointFields(String
    listName, String viewId)    at Microsoft.Samples.SqlServer.SSIS.SharePointUtility.ListServiceUtility.GetFields(Uri sharepointUri, String listName, String viewName)    at Microsoft.Samples.SqlServer.SSIS.SharePointListAdapters.SharePointListSource.GetAccessibleSharePointColumns(String
    sharepointUrl, String listName, String viewName)    at Microsoft.Samples.SqlServer.SSIS.SharePointListAdapters.SharePointListSource.ValidateSharePointColumns()    at Microsoft.Samples.SqlServer.SSIS.SharePointListAdapters.SharePointListSource.Validate()   
    at Microsoft.SqlServer.Dts.Pipeline.ManagedComponentHost.HostValidate(IDTSManagedComponentWrapper90 wrapper).
    Any suggests to me, in order to solve this issue? Many thanks

    Unfortunately your post is off topic as it's not specific to SQL Server Samples and Community Projects.  
    This is a standard response I’ve written in advance to help the many people who post their question in this forum in error, but please don’t ignore it.  The links I provide below will help you determine the right forum to ask your question in.
    For technical issues with Microsoft products that you would run into as an end user, please visit the Microsoft Answers forum ( http://answers.microsoft.com ) which has sections for Windows, Hotmail,
    Office, IE, and other products.
    For Technical issues with Microsoft products that you might have as an IT professional (like technical installation issues, or other IT issues), please head to the TechNet Discussion forums at http://social.technet.microsoft.com/forums/en-us, and
    search for your product name.
    For issues with products you might have as a Developer (like how to talk to APIs, what version of software do what, or other developer issues), please head to the MSDN discussion forums at http://social.msdn.microsoft.com/forums/en-us, and
    search for your product or issue.
    If you’re asking a question particularly about one of the Microsoft Dynamics products, a great place to start is here: http://community.dynamics.com/
    If you think your issue is related to SQL Server Samples and Community Projects and I've flagged it as Off-topic, I apologise.  Please repost your question and include as much detail as possible about your problem so that someone can assist you further. 
    If you really have no idea where to post your question please visit the Where is the forum for…? forum http://social.msdn.microsoft.com/forums/en-us/whatforum/
    When you see answers and helpful posts, please click Vote As Helpful,
    Propose As Answer, and/or Mark As Answer
    Jeff Wharton
    MSysDev (C.Sturt), MDbDsgnMgt (C.Sturt), MCT, MCPD, MCSD, MCSA, MCITP, MCDBA
    Blog: Mr. Wharty's Ramblings
    Twitter: @Mr_Wharty
    MC ID:
    Microsoft Transcript

  • Configure Moss 2007 Doc. Library as a KM Repository

    I have followed the guide that can be found here http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/e0489149-5c93-2a10-8aaa-da1cd0b6edf3
    that shows how to configure the SAP WebDAV Repository Manager to integrate Sharepoint.
    I have managed to complete all of the steps, but errors shown at the end of this post are appearing in the log files, and I am unable to see the repository within the KM Content area.
    Can anyone help me with how to configure this correctly?
    I am using the following versions:
    SAP EP 7 SP15
    MOSS 2007 SP2
    we have implemented single sign of from Windows to EP via kerberos (SPNego), and this functions correctly for users browsing the EP portal - they are not prompted to login at all, so I have not used user mapping.
    If anyone can help me with what to look for in order to get this sorted out, it would be very much appreciated!!
    Thanks in advance
    Regards
    Richard
    Errors in event log on the EP server:
    the repository returned no root collection (null): WebDavRepositoryTest_3
    and also this at about the same time:
    Message : com.sapportals.wcm.repository.AuthorizationRequiredException
    ...... (Please let me know if the full error details will be helpful - I have cut it off so this is not too long...)

    hi,
    we have the same issue.
    did you resoled it ? how ?
    thanks,
    moti

  • Search Crawl status throwing error

    Hi,
    We installed MOSS on our clients machine and Search has been problematic from the getgo.
    When we have performed a forced restart on the osearch service (http://bruteforcesharepoint.blogspot.com/2009_04_01_archive.html), within the next couple of hours the crawler status wil be Error. But, search at this point is still working. Indexing basically
    stops and event errors are thrown in the event logs, but searching is still possible.
    Here are the symptoms:
    the Crawl status is 'Error'
    /ssp/admin/_layouts/searchsspsettings.aspx (Search Settings page) is throwing a HTTP 403 error : The website declined to show this webpage.
    I can still get to the Search Administration, and when I go to view the Content Sources or view the crawl logs specifically, I get the same 403 error.
    When I try to 'Reset all crawled content', that is when Search dies. I get the message an exception error message.
    When I try to stop the Office SharePoint Search service in Central Admin, it hangs in the Stopping state. It'll only come out of this if I go through the task manger and manually force the process to stop and then start it again using the
    stsadm -o osearch -action start -role index
    command then search starts again. And I have to reset the crawled content and perform a full crawl (apart from also reconfiguring the indexer). I can successfully perform a full crawl, takes about 2 hours and it does an incremental crawl for the first
    couple of increments. Then all of a sudden I get the follow errors in the event view log:
    Error event ID: 1030
    Source: Userenv
    Desc: Windows cannot query for the list of Group Policy objects. Check the event log for possible messages previously logged by the policy engine that describes the reason for this.
    And then we get thousands on the following error (the indexing schedule timer job basically stops working):
    Error event id 6398
    Source: Windows SharePoint Services 3
    Desc: The Execute method of job definition Microsoft.Office.Server.Search.Administration.IndexingScheduleJobDefinition (ID b6d1868e-26a5-4ae2-b922-0f129fd77a61) threw an exception. More information is included below.
    Access is denied. (Exception from HRESULT: 0x80070005 (E_ACCESSDENIED))
    Now before you ask me about permission, we have created an admin user called spadmin who has access to everything and is the default content access account as well. It is also spadmin that is running the osearch, not the network account.
    Can anyone provide any insights into why we are experiencing this issue??? We have gone round in circles trying to figure out what we're doing wrong or whether this is a sharepoint issue.
    any help would be much appreciated.
    Cheers
    danielle

    hi danielle,
    I was also in a same sort of situation,  here is what i did:
    1> locate the catalog file directory.
    2>place CI dumps in the 12 hive/bin folder.
    3>run the command
    cidump -dump "catalog filedirectory location" -widset>c:\xyz.txt
    4>search with the word "exception" if there is a hit eg
    error:excetion caught xxxxxxxxx in shadow index xxxxxxxx then u need to delete this index.
    5>run below command
    cidump -removeindex "catalog filedirectory location" -i xxxxxxxxx
    xxxxxxxx signifies the shadow index .
    6> restart the searchservice manually,timersevice.
    if the issue persists let us know
    Moderator Note: NEVER propose your own posts as answers. The function is for proposing the good answers of
    other people. Not for self-proposing.
    (Also don't write let me know - these are forums not private conversations)

  • How to set Current User in moss 2007 for Authentication....

    Hi Team,
    We have current running website where we have user id to login but now requirement is that user will login with his Email address. Email is stored in active directory, I'm able to login with fba authentication.  
    I'm able to retrieve the user id of current user using Email but
    I am NOT able to set that user id for current user for further authentication using Moss 2007.
    Please find the source code for login and authentication page:
    //---------------------------Login Page--------------------//
        public partial class Login : System.Web.UI.UserControl
            #region Decleration
            string LoginStatus = null;
            string UserCurrent = null;
           static SPUser siteuser;
           // private static string _randomNumbercheck;
            //private static bool IsLoggedin;      
            Random random;
            #endregion       
            #region Page Load
            protected void Page_Load(object sender, EventArgs e)
                this.hplchangepw.Visible = false;
                this.hplmytask.Visible = false;
                string desturl = SPContext.Current.Site.Url;
                if (!IsPostBack)
                    try
                        string str;
     SPWeb web = SPContext.Current.Web;
                             SPUser siteuser = web. .CurrentUser; 
                        //SPUser siteuser = spWeb.EnsureUser(username);
                        str = (siteuser == null ? "" : siteuser.ToString());
                        this.Login.TitleText = "";
                        this.Login.DestinationPageUrl = desturl;
                        //if (HttpContext.Current.Request.UrlReferrer != null)
                        //    if (this.Page.User.Identity.IsAuthenticated && ((Session["test1"].ToString() == _randomNumbercheck) || LoginControl.Class1.IsLoggedIn))
                         if (this.Page.User.Identity.IsAuthenticated)
                               // _randomNumbercheck = "";
                                this.hplregister.Visible = false;
                                this.hplforgetpassw.Visible = false;
                                this.hplchangepw.Visible = true;
                                this.hplmytask.Visible = true;
                                this.Login.Visible = false;
      if (str.StartsWith("fba_"))
                                    string name = str.Split(':').GetValue(1).ToString();
                                    LoginnameLbl.Text = "<b>" + name.ToString() + "</b>,<br><br>Haryana Urban Development Authority Welcomes
    You. <br><br><br><b><i>\"In The Service Of Masses\"</i></b>";
                                else
                                    string name = str.Split('\\').GetValue(1).ToString();
                                    LoginnameLbl.Text = "<b>" + name.ToString() + "</b>,<br><br>Haryana Urban Development Authority Welcomes
    You. <br><br><br><b><i>\"In The Service Of Masses\"</i></b>";
                            else
                                this.hplchangepw.Visible = false;
                                this.hplmytask.Visible = false;
                                this.Login.Visible = true;
                                LoginnameLbl.Text = "";
                                HttpContext.Current.Session.Clear();
                                HttpContext.Current.Session.Abandon();
                    catch (Exception ex)
                        //HttpContext.Current.Session.Clear();
                        //HttpContext.Current.Session.Abandon();
            #endregion
            #region Get Audit Informations
            public void GetUserInfo()
                try
                    string user = Login.UserName;
                    SPSecurity.RunWithElevatedPrivileges(new SPSecurity.CodeToRunElevated(delegate()
                        using (SPSite oSite = new SPSite(SPContext.Current.Site.ID))
                            using (SPWeb web1 = oSite.OpenWeb(SPContext.Current.Web.ID))
                                SPListItem item = null;
                                SPListItemCollection listItems = null;
                                web1.AllowUnsafeUpdates = true;
                                siteuser = web1.EnsureUser(Session["User"].ToString());// spWeb.EnsureUser(login);
                                listItems = web1.Lists["Audit_Trail"].Items;
                                item = listItems.Add();
                                item["User_ID"] = UserCurrent.ToString(); //Request.ServerVariables["AUTH_USER"];
                                string ip = Request.ServerVariables["HTTP_X_FORWARDED_FOR"];
                                if ((ip == null) || (ip == "") || (ip.ToLower() == "unknown"))
                                    ip = Request.ServerVariables["REMOTE_ADDR"];
                                item["IP_Address"] = ip;
                                item["Login_Date"] = System.DateTime.Now;
                                item["Login_Status"] = LoginStatus.ToString();
                                item.Update();
                                web1.AllowUnsafeUpdates = false;
                                web1.Dispose();
                                oSite.Dispose();
                catch (Exception ex)
            #endregion
            #region MD5Encryption
            private string MD5Encryption(string strToEncrypt)
                System.Text.UTF8Encoding ue = new System.Text.UTF8Encoding();
                byte[] bytes = ue.GetBytes(strToEncrypt);
                // encrypt bytes
                System.Security.Cryptography.MD5CryptoServiceProvider md5 = new System.Security.Cryptography.MD5CryptoServiceProvider();
                byte[] hashBytes = md5.ComputeHash(bytes);
                // Convert the encrypted bytes back to a string (base 16)
                string hashString = "";
                for (int i = 0; i < hashBytes.Length; i++)
                    hashString += Convert.ToString(hashBytes[i], 16).PadLeft(2, '0');
                return hashString.PadLeft(32, '0');
            #endregion
            #region Generate Random Code
            private string GenerateRandomCode()
                random = new Random();
                string s = "";
                for (int i = 0; i <= 6; i++)
                    s = string.Concat(s, this.random.Next(10).ToString());
               // _randomNumbercheck = s;
                return s;
            #endregion
            #region Generate Random String
            public string GenerateHashKey()
                StringBuilder myStr = new StringBuilder();
                myStr.Append(Request.Browser.Browser);
                myStr.Append(Request.Browser.Platform);
                myStr.Append(Request.Browser.MajorVersion);
                myStr.Append(Request.Browser.MinorVersion);
                myStr.Append(Request.LogonUserIdentity.User.Value);
                SHA1 sha = new SHA1CryptoServiceProvider();
                byte[] hashdata = sha.ComputeHash(Encoding.UTF8.GetBytes(myStr.ToString()));
                return Convert.ToBase64String(hashdata);
            #endregion
            #region Check User Authentication
            protected void Login_Authenticate(object sender, AuthenticateEventArgs e)
                Utility hu=new Utility();
                int CheckValue = 0;
                string user = Login.UserName;
                if(user.Contains("@"))
                    CheckValue=1;
                //Encrypted Password
                string HPassword = hash.Value;
                // Verify that the username/password pair is valid
                SqlConnection con = new SqlConnection();
                SqlCommand cmd = new SqlCommand();
                con.ConnectionString = System.Configuration.ConfigurationSettings.AppSettings["FBAPortalConnection"];
                if (con.State == ConnectionState.Closed)
                    con.Open();
                cmd.CommandType = CommandType.StoredProcedure;
                cmd.CommandText = "sp_Portal_ValidatePortalUser";
                cmd.Connection = con;
                cmd.Parameters.Add("@ID", SqlDbType.NVarChar, 256).Value = Login.UserName;
                cmd.Parameters.Add("@Password", SqlDbType.NVarChar, 300).Value = HPassword.ToString();
                cmd.Parameters.Add("@CheckVal", SqlDbType.Int).Value = CheckValue;
                cmd.Parameters.Add(new SqlParameter("@P_Return", SqlDbType.NVarChar, 256));
                cmd.Parameters["@P_Return"].Direction = ParameterDirection.Output;
                cmd.ExecuteNonQuery();
                string CurrentUserName = cmd.Parameters["@P_Return"].Value.ToString();
                cmd.Dispose();
                con.Close();
                //int x = hu.ValidatePortalUser(Login.UserName, HPassword.ToString(), CheckValue);
                if (CurrentUserName != "0")
                    Page.Session["User"] = CurrentUserName;
                    string UserName = Session["User"].ToString();
                    UserCurrent = UserName;
                    LoginStatus = "Successfull";
                    GetUserInfo();
                    e.Authenticated = true;
                    //Session["CustomAuthKey"] = MD5Encryption(Request.ServerVariables["Remote_Addr"] + Request.ServerVariables["Http_Cookie"] + Request.ServerVariables["Auth_User"]);
                    Session["CustomAuthKey"] = MD5Encryption(Request.ServerVariables["Remote_Addr"] + Request.ServerVariables["Http_Cookie"] + UserCurrent.ToString());
                    FormsAuthenticationTicket tkt;
                    String cookiestr;
                    HttpCookie ck;
                    tkt = new FormsAuthenticationTicket(1, UserName, DateTime.Now, DateTime.Now.AddMinutes(15), false, GenerateHashKey());
                    cookiestr = FormsAuthentication.Encrypt(tkt);
                    ck = new HttpCookie(FormsAuthentication.FormsCookieName, cookiestr);
                    ck.Path = FormsAuthentication.FormsCookiePath;
                    Response.Cookies.Add(ck);
                else
                    // Username/password are not valid...
                    Login.FailureText = "Invalid Username / password";
                    e.Authenticated = false;
            #endregion
            #region Login Error
            protected void Login_LoginError(object sender, EventArgs e)
                // Determine why the user could not login
                Login.FailureText = "Your login attempt was not successful. Please try again.";
                // Does there exist a User account for this user
                MembershipUser usrInfo = Membership.GetUser(Login.UserName);
                if (usrInfo != null)
                    // Is this user locked out?
                    if (usrInfo.IsLockedOut)
                        Login.FailureText = "Your account has been locked out because of too many invalid login attempts. Please contact the administrator to have your account unlocked.";
                    else if (!usrInfo.IsApproved)
                        Login.FailureText = "Your account has not yet been approved. You cannot login until an administrator has approved your account.";
                    else
                        Login.FailureText = "Invalid UserName / Password";
                        UserCurrent = usrInfo.UserName;
                        LoginStatus = "UnSucessfull";
                        GetUserInfo();
                else
                    Login.FailureText = "Invalid UserName / Password";
            #endregion
            #region Logged In
            protected void Login_LoggedIn(object sender, EventArgs e)
                string UserName = Session["User"].ToString();           
                LoginControl.Class1.IsLoggedIn = true;
                //If User Email is [email protected] it will be invalid. Prompt the user to update the email id. 
                LoginControl.Utility hu = new LoginControl.Utility();
                DataSet dsuser = hu.getUserDetails(UserName);
                string email;
                if (dsuser.Tables[0].Rows.Count > 0)
                    email = dsuser.Tables[0].Rows[0]["Email"].ToString();
                    if (email.Equals("[email protected]"))
                        Response.Redirect("/Pages/UserUpdates.aspx", true);
                    else
            #endregion
    //-------------------------------- Authentication code----------------------------//
      private void authenticateuser()
                try
                    if (SPContext.Current.Web.CurrentUser != null)
                        using (SPSite spSite = new SPSite(SPContext.Current.Site.Url))
                            using (SPWeb spWep = spSite.OpenWeb())
                                Utility hu = new Utility();
    _userid = spWep.CurrentUser.Name;
                                DataSet ds = new DataSet();
                                ds = hu.GetPlotid(_userid);
                                if (ds.Tables[0].Rows.Count > 0)
                                    _plotid = ds.Tables[0].Rows[0]["plotid"].ToString();
                                else
                                    if (!SPContext.Current.Web.UserIsSiteAdmin && !SPContext.Current.Web.UserIsWebAdmin)
                                        if (!SPContext.Current.Web.IsCurrentUserMemberOfGroup(SPContext.Current.Web.Groups["Allottee"].ID))
                                            SPUtility.HandleAccessDenied(new Exception("You do not have access to this page"));
                    else
                    { SPUtility.HandleAccessDenied(new Exception("Please login")); }
                catch (Exception ex)
                    //  lblMessage.Text = "Error:" + ex.Message;
                    ShowMessage("Following error has occured while executing the desried event :- " + ex.Message);
    Mohan Prakash

    1. Current work flow :-In web site the user have to register themselves in the web site and enters his details along with user id, password and email id. Once user
    is registered then he will login with his user id and password. The user id is picked from login control and that set in “SPWeb.CurrentUser” as user id and system uses "Membership.ValidateUser" method to Authenticate user for login. 
    2. New Requirement: we would like to facilitate the user to login with Email id as well as user id. 
    Problem: 
    We replaced "Membership.ValidateUser" method to our own method to Authenticate user with email id/user id and password. 
    When user is login with user id and password it is working successfully but in the case of email id and password –“the email id is picked from user control and set as
    "HttpContext.Current.User.Identity.Name" but we are not getting “SPWeb.CurrentUser” and it shows null value.” 
    We are able to get user id from database using email id of user. Please help us how we can set user id in "SPContext.Current.Web.CurrentUser".
    Mohan Prakash

  • Search not working in site collections migrated from MOSS 2007 to SharePoint Server 2013

    I have few site collections migrated from MOSS 2007 to SharePoint Server 2013 using database attach (first to 2010 and then to 2013).
    When I search for anything in these migrated sites, I get "Nothing here matches your search".
    Checked the crawl log for these sites and see this error for all items.
    The content processing component failed to process the security descriptor of the item. ( The input ACL is invalid (1210 / 2072).; ; SearchID =
    Some-Random-GUID )
    Search works fine in newly created sites in 2013 environment.
    Please help.

    Hello
    Did you migrate your users to claims when you migrated the databases?
    $wa = Get-SPWebApplication http://yourdom.com
    $wa.MigrateUsers($true)
    $wa.ProvisionGlobally()
    I assume the sites that work and the migrated content are in the same database?
    MCITP-EA | "Never test how deep the water is with both feet"

  • After upgrade SP-Crawl Error: The SharePoint item being crawled returned an error when attempting to download the item.

    Hi All - After the upgrade, I am getting SP-Crawl Error for certain links. I check the Crawl component has proper permission.
    Google is showing some article like
    http://blog.karstein-consulting.com/2012/04/20/error-in-crawl-log-the-sharepoint-item-being-crawled-returned-an-error-when-attempting-to-download-the-item/
    not sure if this resolution is referring to 2010 and/or 2013. 
    I checked the registery editor. I couldn't find 14.0 under the Office Server.
    Any clue?
    Regards,
    Khushi
    Khushi

    I checked the web application policy the search crawl account has full read permission.
    Crawl
    Fiddler
    Log Error referring the Correlation ID
    01/06/2014 13:05:06.14  w3wp.exe (0x1698)                        0x0118 SharePoint Foundation        
     Monitoring                     nasq Medium   Entering monitored scope (Request (GET:/sites/HR/Shared%20Documents/Benefits/Insurance%20Benefits/Life%20Insurance/Basic%20Life%20and%20ADD)).
    Parent No 
    01/06/2014 13:05:06.14  w3wp.exe (0x1698)                        0x0118 SharePoint Foundation        
     Logging Correlation Data       xmnv Medium   Name=Request (GET:<SiteURL>/sites/HR/Shared%20Documents/Benefits/Insurance%20Benefits/Life%20Insurance/Basic%20Life%20and%20ADD) e8b1679c-0476-70d4-9fcd-2cef5be44461
    01/06/2014 13:05:06.14  w3wp.exe (0x1698)                        0x0118 SharePoint Foundation        
     Authentication Authorization   agb9s Medium   Non-OAuth request. IsAuthenticated=True, UserIdentityName=, ClaimsCount=0 e8b1679c-0476-70d4-9fcd-2cef5be44461
    01/06/2014 13:05:06.15  w3wp.exe (0x1698)                        0x1738 SharePoint Foundation        
     General                        af71 Medium   HTTP Request method: GET e8b1679c-0476-70d4-9fcd-2cef5be44461
    01/06/2014 13:05:06.15  w3wp.exe (0x1698)                        0x1738 SharePoint Foundation        
     General                        af75 Medium   Overridden HTTP request method: GET e8b1679c-0476-70d4-9fcd-2cef5be44461
    01/06/2014 13:05:06.15  w3wp.exe (0x1698)                        0x1738 SharePoint Foundation        
     General                        af74 Medium   HTTP request URL: /sites/HR/Shared%20Documents/Benefits/Insurance%20Benefits/Life%20Insurance/Basic%20Life%20and%20ADD e8b1679c-0476-70d4-9fcd-2cef5be44461
    01/06/2014 13:05:06.17  w3wp.exe (0x1698)                        0x1738 SharePoint Foundation        
     Files                          aise3 Medium   Failure when fetching document. 0x80070090 e8b1679c-0476-70d4-9fcd-2cef5be44461
    01/06/2014 13:05:06.17  w3wp.exe (0x1698)                        0x1960 SharePoint Foundation        
     Monitoring                     b4ly Medium   Leaving Monitored Scope (Request (GET:<SiteURL>/sites/HR/Shared%20Documents/Benefits/Insurance%20Benefits/Life%20Insurance/Basic%20Life%20and%20ADD)).
    Execution Time=20.5461867360237 e8b1679c-0476-70d4-9fcd-2cef5be44461
    01/06/2014 13:05:06.17  w3wp.exe (0x1698)                        0x1960 SharePoint Foundation        
     Monitoring                     b4ly Medium   Leaving Monitored Scope (Request (GET:<SiteURL>/sites/HR/Shared%20Documents/Benefits/Insurance%20Benefits/Life%20Insurance/Basic%20Life%20and%20ADD)).
    Execution Time=29.917489513332 e8b1679c-0476-70d4-9fcd-2549ba3ee9d4
    01/06/2014 13:05:06.17  w3wp.exe (0x1698)                        0x1960 SharePoint Foundation        
     Monitoring                     nasq Medium   Entering monitored scope (Request (GET:<SiteURL>nsurance%20Benefits%2fLife%20Insurance%2fBasic%20Life%20and%20ADD&FolderCTID=0x01200039DA632EEACF264685CF39D68A18F7C8)).
    Parent No 
    Any clue?
    Regards,
    Khushi
    Khushi

  • Full crawl vs incremental crawl

    Hi All,
    I do understand what the full crawling is, can anyone please tell me what are all things that  Search Service will crawl in case of incremental crawling.
    Regards Amit K

    Here some more information
    Full Crawl: “Full crawl” crawls entire content under a content source (depending upon two settings specified at the time of creating the content source. These settings are “Content Source type” and “Crawl Settings”).
    Incremental Crawl: “Incremental crawl” crawls the content which has been added/modified after last successful crawl.
    Why do we need incremental Crawl?
    Though “Full crawl” crawls every bit and piece of content under a content source but we surely need Incremental crawl as it crawls content which has been added/modified after last successful crawl.
    Full crawls will take more time and resource to complete than Incremental crawls. You should consider following facts before going for full crawl instead of incremental crawl.
    1.As compared with incremental crawls, full crawls chew up more memory and CPU cycles on the index .
    2.Full crawls consume more memory and CPU cycles on the Web Front End servers when crawling content in your farm.
    3.Full crawls use more network bandwidth than incremental crawls.
    Crawling puts an overhead on resources. If some content is already been crawled and indexed, why do we need to crawl it again? Therefore incremental crawl is used in such cases to take care of any added/modified content after last successful crawl.
    There are some scenarios where incremental crawl doesn’t work and you need to run full crawl.
    Why do we need Full Crawl?
    1. Software updates or service packs installation on servers in the farm.
    2. When an SSP administrator added new managed property.
    3. Crawl rules have been added, deleted, or modified.
    4. Full crawl is required to repair corrupted index. In this case, system may attempt a full crawl (depending on severity of corruption)
    5. A full crawl of the site has never been done.
    6. To detect security changes those were made on file shares after the last full crawl of the file share.
    7. In case, incremental crawl is failing consecutively. In rare cases, if an incremental crawl fails one hundred consecutive times at any level in a repository, the index server removes the affected content from the index.
    8. To reindex ASPX pages on Windows SharePoint Services 3.0 or Office SharePoint Server 2007 sites. The crawler cannot discover when ASPX pages on Windows SharePoint Services 3.0 or MOSS sites have changed. Because of this, incremental crawls do not reindex
    views or home pages when individual list items are deleted.
    The system does a full crawl even when an incremental crawl is requested under the following circumstances:
    · A shared services administrator stopped the previous crawl.
    · A content database was restored. This applies to MOSS and Windows SharePoint Services 3.0 content databases only.
    source :
    http://sannadisarath.blogspot.com/2011/03/incremental-crawl-vs-full-crawl.html
    My SharePoint Blog

  • Sharepoint crawl db growing exceptionally after full crawl run..

    while we run the crawl sharepoint crawl db growing exceptionally.....drive is running out of space.
    lots of errors coming in crawl.....it is trying to do the crawl even on thoese sites which were moved, deleted or URL changed.
    everything is showing in crawl.....any script/suggestion please?

    how many documents you are crawling? full crawl will crawl all sites if its deleted then it delete from index but 1st check and make sure.
    check these post:
    http://social.technet.microsoft.com/Forums/sharepoint/en-US/f0d95c9c-5797-4818-9189-609a48af59d0/crawl-database-size-is-growing-rapidly?forum=sharepointgeneralprevious
    http://www.itwalkthru.com/2011/11/sharepoint-2010-application-crawl-store.html
    for deletion of item from index may be need to play with deletion policy:
    http://technet.microsoft.com/en-us/library/hh127009.aspx
    Please remember to mark your question as answered &Vote helpful,if this solves/helps your problem. ****************************************************************************************** Thanks -WS MCITP(SharePoint 2010, 2013) Blog: http://wscheema.com/blog

  • Showing error while creating IT0000 ee record

    Hi,
    I am new to HR functional. Please do need full.
    When i am creating record in infotype 0000, it is showing error like No entry in table T503 for     779.
    I am giving value 7 fro employee group and 79 for employee subgroup.
    Is there any where that i can enter value 779.
    Thanks,
    Thrilleswar

    Hi ram,
    Thanks so much..
    this error is resolved but the other error is showing like Number range 01 is internal, do not enter a personnel number.
    Is there any place that i can be able to make my number work here.
    Thanks,
    Thrilleswar

Maybe you are looking for