Sp crawl db size

our Crawl Database Size is growing rapidly and drive is having only 17 GB ramaining...what is side effect of it.
Is there will be any impact since drive is shrinking rapidly...how much must be available for crawl db?

Check this link. will help you.
http://technet.microsoft.com/en-us/library/cc262531(v=office.12).aspx
Sugumaran Srinuvasan

Similar Messages

  • Crawling the Web with Java

    Hi everyone.
    I've been trying to learn how to make a web crawler in java following this detailed albeit old tutorial: http://www.devarticles.com/c/a/Java/Crawling-the-Web-with-Java/
    The SearchCrawler class they feature can be found in two parts:
    First part: http://www.devarticles.com/c/a/Java/Crawling-the-Web-with-Java/3/
    Second part: http://www.devarticles.com/c/a/Java/Crawling-the-Web-with-Java/4/
    I don't want to copy and paste the code because it is really long and an eyesore if viewing here.
    I get a lot of errors when compiling of which I do not understand. The majority of the errors (62 of them to be precise) are "class, interface or enum expected" errors, with the remaining few being "illegal start of type" and "<identifier> expected" errors.
    Can someone here perhaps take a look at it and compile it and see if they also get the same errors? I realise it is an old tutorial but there are hardly any detailed resources I can find for java web crawlers.
    Thanks.

    Odd I can't seem to log into my account. Never mind.
    I have used java before, the problem here I suppose is I'm not good enough to spot what's the problem. The code seems fine bracket wise and it has really left me stumped.
    If someone code put it in their editor and attempt to compile and see if I'm not the only one that's having a problem.. that would be much appreciated.
    For your convenience... the code from the example I linked:
    import java.awt.*;
    import java.awt.event.*;
    import java.io.*;
    import java.net.*;
    import java.util.*;
    import java.util.regex.*;
    import javax.swing.*;
    import javax.swing.table.*;
    // The Search Web Crawler
    public class SearchCrawler extends JFrame
      // Max URLs drop-down values.
      private static final String[] MAX_URLS =
        {"50", "100", "500", "1000"};
      // Cache of robot disallow lists.
      private HashMap disallowListCache = new HashMap();
      // Search GUI controls.
      private JTextField startTextField;
      private JComboBox maxComboBox;
      private JCheckBox limitCheckBox;
      private JTextField logTextField;
      private JTextField searchTextField;
      private JCheckBox caseCheckBox;
      private JButton searchButton;
      // Search stats GUI controls.
      private JLabel crawlingLabel2;
      private JLabel crawledLabel2;
      private JLabel toCrawlLabel2;
      private JProgressBar progressBar;
      private JLabel matchesLabel2;
      // Table listing search matches.
      private JTable table;
      // Flag for whether or not crawling is underway.
      private boolean crawling;
      // Matches log file print writer.
      private PrintWriter logFileWriter;
      // Constructor for Search Web Crawler.
      public SearchCrawler()
        // Set application title.
        setTitle("Search Crawler");
        // Set window size.
        setSize(600, 600);
         // Handle window closing events.
        addWindowListener(new WindowAdapter() {
         public void windowClosing(WindowEvent e) {
           actionExit();
        // Set up File menu.
        JMenuBar menuBar = new JMenuBar();
        JMenu fileMenu = new JMenu("File"); 
        fileMenu.setMnemonic(KeyEvent.VK_F);
        JMenuItem fileExitMenuItem = new JMenuItem("Exit",
          KeyEvent.VK_X);
        fileExitMenuItem.addActionListener(new ActionListener() {
          public void actionPerformed(ActionEvent e) { 
            actionExit();
        fileMenu.add(fileExitMenuItem);
        menuBar.add(fileMenu);
        setJMenuBar(menuBar);
        // Set up search panel.
        JPanel searchPanel = new JPanel();
        GridBagConstraints constraints;
        GridBagLayout layout = new GridBagLayout();
        searchPanel.setLayout(layout);
        JLabel startLabel = new JLabel("Start URL:");
        constraints = new GridBagConstraints();
        constraints.anchor = GridBagConstraints.EAST; 
        constraints.insets = new Insets(5, 5, 0, 0);
        layout.setConstraints(startLabel, constraints);
        searchPanel.add(startLabel);
        startTextField = new JTextField();
        constraints = new GridBagConstraints();
        constraints.fill = GridBagConstraints.HORIZONTAL;
        constraints.gridwidth = GridBagConstraints.REMAINDER;
        constraints.insets = new Insets(5, 5, 0, 5);
        layout.setConstraints(startTextField, constraints);
        searchPanel.add(startTextField);
        JLabel maxLabel = new JLabel("Max URLs to Crawl:");
        constraints = new GridBagConstraints();
        constraints.anchor = GridBagConstraints.EAST;
        constraints.insets = new Insets(5, 5, 0, 0);
        layout.setConstraints(maxLabel, constraints);
        searchPanel.add(maxLabel);
        maxComboBox = new JComboBox(MAX_URLS);
        maxComboBox.setEditable(true);
        constraints = new GridBagConstraints();
        constraints.insets = new Insets(5, 5, 0, 0);
        layout.setConstraints(maxComboBox, constraints);
        searchPanel.add(maxComboBox);
        limitCheckBox =
          new JCheckBox("Limit crawling to Start URL site");
        constraints = new GridBagConstraints();
        constraints.anchor = GridBagConstraints.WEST;
        constraints.insets = new Insets(0, 10, 0, 0);
        layout.setConstraints(limitCheckBox, constraints);
        searchPanel.add(limitCheckBox);
        JLabel blankLabel = new JLabel();
        constraints = new GridBagConstraints();
        constraints.gridwidth = GridBagConstraints.REMAINDER;
        layout.setConstraints(blankLabel, constraints);
        searchPanel.add(blankLabel);
        JLabel logLabel = new JLabel("Matches Log File:");
        constraints = new GridBagConstraints();
        constraints.anchor = GridBagConstraints.EAST;
        constraints.insets = new Insets(5, 5, 0, 0);
        layout.setConstraints(logLabel, constraints);
        searchPanel.add(logLabel);
        String file =
          System.getProperty("user.dir") +
          System.getProperty("file.separator") +
          "crawler.log";
        logTextField = new JTextField(file);
        constraints = new GridBagConstraints();
        constraints.fill = GridBagConstraints.HORIZONTAL;
        constraints.gridwidth = GridBagConstraints.REMAINDER;
        constraints.insets = new Insets(5, 5, 0, 5);
        layout.setConstraints(logTextField, constraints);
        searchPanel.add(logTextField);
        JLabel searchLabel = new JLabel("Search String:");
        constraints = new GridBagConstraints();
        constraints.anchor = GridBagConstraints.EAST; 
        constraints.insets = new Insets(5, 5, 0, 0);
        layout.setConstraints(searchLabel, constraints);
        searchPanel.add(searchLabel);
        searchTextField = new JTextField();
        constraints = new GridBagConstraints();
        constraints.fill = GridBagConstraints.HORIZONTAL;
        constraints.insets = new Insets(5, 5, 0, 0);
        constraints.gridwidth= 2;
        constraints.weightx = 1.0d;
        layout.setConstraints(searchTextField, constraints);
        searchPanel.add(searchTextField);
        caseCheckBox = new JCheckBox("Case Sensitive");
        constraints = new GridBagConstraints();
        constraints.insets = new Insets(5, 5, 0, 5);
        constraints.gridwidth = GridBagConstraints.REMAINDER;
        layout.setConstraints(caseCheckBox, constraints);
        searchPanel.add(caseCheckBox);
        searchButton = new JButton("Search");
        searchButton.addActionListener(new ActionListener() {
          public void actionPerformed(ActionEvent e) {
            actionSearch();
        constraints = new GridBagConstraints();
        constraints.gridwidth = GridBagConstraints.REMAINDER;
        constraints.insets = new Insets(5, 5, 5, 5);
        layout.setConstraints(searchButton, constraints);
        searchPanel.add(searchButton);
        JSeparator separator = new JSeparator();
        constraints = new GridBagConstraints();
        constraints.fill = GridBagConstraints.HORIZONTAL;
        constraints.gridwidth = GridBagConstraints.REMAINDER;
        constraints.insets = new Insets(5, 5, 5, 5);
        layout.setConstraints(separator, constraints);
        searchPanel.add(separator);
        JLabel crawlingLabel1 = new JLabel("Crawling:");
        constraints = new GridBagConstraints();
        constraints.anchor = GridBagConstraints.EAST;
        constraints.insets = new Insets(5, 5, 0, 0);
        layout.setConstraints(crawlingLabel1, constraints);
        searchPanel.add(crawlingLabel1);
        crawlingLabel2 = new JLabel();
        crawlingLabel2.setFont(
          crawlingLabel2.getFont().deriveFont(Font.PLAIN));
        constraints = new GridBagConstraints();
        constraints.fill = GridBagConstraints.HORIZONTAL;
        constraints.gridwidth = GridBagConstraints.REMAINDER;
        constraints.insets = new Insets(5, 5, 0, 5);
        layout.setConstraints(crawlingLabel2, constraints);
        searchPanel.add(crawlingLabel2);
        JLabel crawledLabel1 = new JLabel("Crawled URLs:");
        constraints = new GridBagConstraints();
        constraints.anchor = GridBagConstraints.EAST;
        constraints.insets = new Insets(5, 5, 0, 0);
        layout.setConstraints(crawledLabel1, constraints);
        searchPanel.add(crawledLabel1);
        crawledLabel2 = new JLabel();
        crawledLabel2.setFont(
          crawledLabel2.getFont().deriveFont(Font.PLAIN));
        constraints = new GridBagConstraints();
        constraints.fill = GridBagConstraints.HORIZONTAL;
        constraints.gridwidth = GridBagConstraints.REMAINDER;
        constraints.insets = new Insets(5, 5, 0, 5);
        layout.setConstraints(crawledLabel2, constraints);
        searchPanel.add(crawledLabel2);
        JLabel toCrawlLabel1 = new JLabel("URLs to Crawl:");
        constraints = new GridBagConstraints();
        constraints.anchor = GridBagConstraints.EAST;
        constraints.insets = new Insets(5, 5, 0, 0);
        layout.setConstraints(toCrawlLabel1, constraints);
        searchPanel.add(toCrawlLabel1);
        toCrawlLabel2 = new JLabel();
        toCrawlLabel2.setFont(
          toCrawlLabel2.getFont().deriveFont(Font.PLAIN));
        constraints = new GridBagConstraints();
        constraints.fill = GridBagConstraints.HORIZONTAL;
        constraints.gridwidth = GridBagConstraints.REMAINDER;
        constraints.insets = new Insets(5, 5, 0, 5);
        layout.setConstraints(toCrawlLabel2, constraints);
        searchPanel.add(toCrawlLabel2);
        JLabel progressLabel = new JLabel("Crawling Progress:");
        constraints = new GridBagConstraints();
        constraints.anchor = GridBagConstraints.EAST;
        constraints.insets = new Insets(5, 5, 0, 0);
        layout.setConstraints(progressLabel, constraints);
        searchPanel.add(progressLabel);
        progressBar = new JProgressBar();
        progressBar.setMinimum(0);
        progressBar.setStringPainted(true);
        constraints = new GridBagConstraints();
        constraints.fill = GridBagConstraints.HORIZONTAL;
        constraints.gridwidth = GridBagConstraints.REMAINDER;
        constraints.insets = new Insets(5, 5, 0, 5);
        layout.setConstraints(progressBar, constraints);
        searchPanel.add(progressBar);
        JLabel matchesLabel1 = new JLabel("Search Matches:");
        constraints = new GridBagConstraints();
        constraints.anchor = GridBagConstraints.EAST;
        constraints.insets = new Insets(5, 5, 10, 0);
        layout.setConstraints(matchesLabel1, constraints);
        searchPanel.add(matchesLabel1);
        matchesLabel2 = new JLabel();
        matchesLabel2.setFont(
          matchesLabel2.getFont().deriveFont(Font.PLAIN));
        constraints = new GridBagConstraints();
        constraints.fill = GridBagConstraints.HORIZONTAL;
        constraints.gridwidth = GridBagConstraints.REMAINDER;
        constraints.insets = new Insets(5, 5, 10, 5);
        layout.setConstraints(matchesLabel2, constraints);
        searchPanel.add(matchesLabel2);
        // Set up matches table.
        table =
          new JTable(new DefaultTableModel(new Object[][]{},
            new String[]{"URL"}) {
          public boolean isCellEditable(int row, int column)
            return false;
        // Set up Matches panel.
        JPanel matchesPanel = new JPanel();
        matchesPanel.setBorder(
          BorderFactory.createTitledBorder("Matches"));
        matchesPanel.setLayout(new BorderLayout());
        matchesPanel.add(new JScrollPane(table),
          BorderLayout.CENTER);
        // Add panels to display.
        getContentPane().setLayout(new BorderLayout());
        getContentPane().add(searchPanel, BorderLayout.NORTH);
        getContentPane().add(matchesPanel,BorderLayout.CENTER);
      // Exit this program.
      private void actionExit() {
        System.exit(0);
      // Handle Search/Stop button being clicked.
      private void actionSearch() {
        // If stop button clicked, turn crawling flag off.
        if (crawling) {
          crawling = false;
          return;
      ArrayList errorList = new ArrayList();
      // Validate that start URL has been entered.
      String startUrl = startTextField.getText().trim();
      if (startUrl.length() < 1) {
        errorList.add("Missing Start URL.");
      // Verify start URL.
      else if (verifyUrl(startUrl) == null) {
        errorList.add("Invalid Start URL.");
      // Validate that Max URLs is either empty or is a number.
      int maxUrls = 0;
      String max = ((String) maxComboBox.getSelectedItem()).trim();
      if (max.length() > 0) {
        try {
          maxUrls = Integer.parseInt(max);
        } catch (NumberFormatException e) {
        if (maxUrls < 1) {
          errorList.add("Invalid Max URLs value.");
      // Validate that matches log file has been entered.
      String logFile = logTextField.getText().trim();
      if (logFile.length() < 1) {
        errorList.add("Missing Matches Log File.");
      // Validate that search string has been entered.
      String searchString = searchTextField.getText().trim();
      if (searchString.length() < 1) {
        errorList.add("Missing Search String.");
      // Show errors, if any, and return.
      if (errorList.size() > 0) {
        StringBuffer message = new StringBuffer();
        // Concatenate errors into single message.
        for (int i = 0; i < errorList.size(); i++) {
          message.append(errorList.get(i));
          if (i + 1 < errorList.size()) {
            message.append("\n");
        showError(message.toString());
        return;
      // Remove "www" from start URL if present.
      startUrl = removeWwwFromUrl(startUrl);
      // Start the Search Crawler.
      search(logFile, startUrl, maxUrls, searchString);
    private void search(final String logFile, final String startUrl,
      final int maxUrls, final String searchString)
      // Start the search in a new thread.
      Thread thread = new Thread(new Runnable() {
        public void run() {
          // Show hour glass cursor while crawling is under way.
          setCursor(Cursor.getPredefinedCursor(Cursor.WAIT_CURSOR));
          // Disable search controls.
          startTextField.setEnabled(false);
          maxComboBox.setEnabled(false);
          limitCheckBox.setEnabled(false);
          logTextField.setEnabled(false);
          searchTextField.setEnabled(false);
          caseCheckBox.setEnabled(false);
          // Switch Search button to "Stop."
          searchButton.setText("Stop");
          // Reset stats.
          table.setModel(new DefaultTableModel(new Object[][]{},
            new String[]{"URL"}) {
            public boolean isCellEditable(int row, int column)
              return false;
           updateStats(startUrl, 0, 0, maxUrls);
          // Open matches log file.
          try {
            logFileWriter = new PrintWriter(new FileWriter(logFile));
          } catch (Exception e) {
            showError("Unable to open matches log file.");
            return;
          // Turn crawling flag on.
          crawling = true;
          // Perform the actual crawling.
          crawl(startUrl, maxUrls, limitCheckBox.isSelected(),
            searchString, caseCheckBox.isSelected());
          // Turn crawling flag off.
          crawling = false;
          // Close matches log file.
          try {
            logFileWriter.close();
          } catch (Exception e) {
            showError("Unable to close matches log file.");
          // Mark search as done.
          crawlingLabel2.setText("Done");
          // Enable search controls.
          startTextField.setEnabled(true);
          maxComboBox.setEnabled(true);
          limitCheckBox.setEnabled(true);
          logTextField.setEnabled(true);
          searchTextField.setEnabled(true);
          caseCheckBox.setEnabled(true);
          // Switch search button back to "Search."
          searchButton.setText("Search");
          // Return to default cursor.
          setCursor(Cursor.getDefaultCursor());
          // Show message if search string not found.
          if (table.getRowCount() == 0) {
            JOptionPane.showMessageDialog(SearchCrawler.this,
              "Your Search String was not found. Please try another.",
              "Search String Not Found",
              JOptionPane.WARNING_MESSAGE);
      thread.start();
    // Show dialog box with error message.
    private void showError(String message) {
      JOptionPane.showMessageDialog(this, message, "Error",
        JOptionPane.ERROR_MESSAGE);
    // Update crawling stats.
    private void updateStats(
      String crawling, int crawled, int toCrawl, int maxUrls)
      crawlingLabel2.setText(crawling);
      crawledLabel2.setText("" + crawled);
      toCrawlLabel2.setText("" + toCrawl);
      // Update progress bar.
      if (maxUrls == -1) {
        progressBar.setMaximum(crawled + toCrawl);
      } else {
        progressBar.setMaximum(maxUrls);
      progressBar.setValue(crawled);
      matchesLabel2.setText("" + table.getRowCount());
    // Add match to matches table and log file.
    private void addMatch(String url) {
      // Add URL to matches table.
      DefaultTableModel model =
        (DefaultTableModel) table.getModel();
      model.addRow(new Object[]{url});
      // Add URL to matches log file.
      try {
        logFileWriter.println(url);
      } catch (Exception e) {
        showError("Unable to log match.");
    // Verify URL format.
    private URL verifyUrl(String url) {
      // Only allow HTTP URLs.
      if (!url.toLowerCase().startsWith("http://"))
        return null;
      // Verify format of URL.
      URL verifiedUrl = null;
      try {
        verifiedUrl = new URL(url);
      } catch (Exception e) {
        return null;
      return verifiedUrl;
    // Check if robot is allowed to access the given URL. private boolean isRobotAllowed(URL urlToCheck) {
      String host = urlToCheck.getHost().toLowerCase();
      // Retrieve host's disallow list from cache.
      ArrayList disallowList =
        (ArrayList) disallowListCache.get(host);
      // If list is not in the cache, download and cache it.
      if (disallowList == null) {
        disallowList = new ArrayList();
        try {
          URL robotsFileUrl =
            new URL("http://" + host + "/robots.txt");
          // Open connection to robot file URL for reading.
          BufferedReader reader =
            new BufferedReader(new InputStreamReader(
              robotsFileUrl.openStream()));
          // Read robot file, creating list of disallowed paths.
          String line;
          while ((line = reader.readLine()) != null) {
            if (line.indexOf("Disallow:") == 0) {
              String disallowPath =
                line.substring("Disallow:".length());
              // Check disallow path for comments and remove if present.
              int commentIndex = disallowPath.indexOf("#");
              if (commentIndex != -1) {
                disallowPath =
                  disallowPath.substring(0, commentIndex);
              // Remove leading or trailing spaces from disallow path.
              disallowPath = disallowPath.trim();
              // Add disallow path to list.
              disallowList.add(disallowPath);
          // Add new disallow list to cache.
          disallowListCache.put(host, disallowList);
        catch (Exception e) {
          /* Assume robot is allowed since an exception
             is thrown if the robot file doesn't exist. */
          return true;
      /* Loop through disallow list to see if
         crawling is allowed for the given URL. */
      String file = urlToCheck.getFile();
      for (int i = 0; i < disallowList.size(); i++) {
        String disallow = (String) disallowList.get(i);
        if (file.startsWith(disallow)) {
          return false;
      return true;
    // Download page at given URL.
    private String downloadPage(URL pageUrl) {
      try {
        // Open connection to URL for reading.
        BufferedReader reader =
          new BufferedReader(new InputStreamReader(
            pageUrl.openStream()));
        // Read page into buffer.
        String line;
        StringBuffer pageBuffer = new StringBuffer();
        while ((line = reader.readLine()) != null) {
          pageBuffer.append(line);
        return pageBuffer.toString();
      } catch (Exception e) {
      return null;
    // Remove leading "www" from a URL's host if present.
    private String removeWwwFromUrl(String url) {
      int index = url.indexOf("://www.");
      if (index != -1) {
        return url.substring(0, index + 3) +
          url.substring(index + 7);
      return (url);
    // Parse through page contents and retrieve links.
    private ArrayList retrieveLinks(
      URL pageUrl, String pageContents, HashSet crawledList,
      boolean limitHost)
      // Compile link matching pattern.
      Pattern p =
        Pattern.compile("<a\\s+href\\s*=\\s*\"?(.*?)[\"|>]",
          Pattern.CASE_INSENSITIVE);
      Matcher m = p.matcher(pageContents);
      // Create list of link matches.
      ArrayList linkList = new ArrayList();
      while (m.find()) {
        String link = m.group(1).trim();
        // Skip empty links.
        if (link.length() < 1) {
          continue;
        // Skip links that are just page anchors.
        if (link.charAt(0) == '#') {
          continue;
        // Skip mailto links.
        if (link.indexOf("mailto:") != -1) {
          continue;
        // Skip JavaScript links.
        if (link.toLowerCase().indexOf("javascript") != -1) {
          continue;
        // Prefix absolute and relative URLs if necessary.
        if (link.indexOf("://") == -1) {
          // Handle absolute URLs.
          if (link.charAt(0) == '/') {
            link = "http://" + pageUrl.getHost() + link;
          // Handle relative URLs.
          } else {
            String file = pageUrl.getFile();
            if (file.indexOf('/') == -1) {
              link = "http://" + pageUrl.getHost() + "/" + link;
            } else {
              String path =
                file.substring(0, file.lastIndexOf('/') + 1);
              link = "http://" + pageUrl.getHost() + path + link;
        // Remove anchors from link.
        int index = link.indexOf('#');
        if (index != -1) {
          link = link.substring(0, index);
        // Remove leading "www" from URL's host if present.
        link = removeWwwFromUrl(link);
        // Verify link and skip if invalid.
        URL verifiedLink = verifyUrl(link);
        if (verifiedLink == null) {
          continue;
        /* If specified, limit links to those
          having the same host as the start URL. */
        if (limitHost &&
            !pageUrl.getHost().toLowerCase().equals(
              verifiedLink.getHost().toLowerCase())) 
          continue;
        // Skip link if it has already been crawled.
        if (crawledList.contains(link)) {
          continue;
        // Add link to list.
        linkList.add(link);
      return (linkList);
    /* Determine whether or not search string is
       matched in the given page contents. */
    private boolean searchStringMatches(
      String pageContents, String searchString,
      boolean caseSensitive)
      String searchContents = pageContents;
      /* If case-sensitive search, lowercase
         page contents for comparison. */
      if (!caseSensitive) {
        searchContents = pageContents.toLowerCase();
      // Split search string into individual terms.
      Pattern p = Pattern.compile("[\\s]+");
      String[] terms = p.split(searchString);
      // Check to see if each term matches.
      for (int i = 0; i < terms.length; i++) {
        if (caseSensitive) {
          if (searchContents.indexOf(terms) == -1) {
    return false;
    } else {
    if (searchContents.indexOf(terms[i].toLowerCase()) == -1) {
    return false;
    return true;
    // Perform the actual crawling, searching for the search string.
    public void crawl(
    String startUrl, int maxUrls, boolean limitHost,
    String searchString, boolean caseSensitive)
    // Set up crawl lists.
    HashSet crawledList = new HashSet();
    LinkedHashSet toCrawlList = new LinkedHashSet();
    // Add start URL to the to crawl list.
    toCrawlList.add(startUrl);
    /* Perform actual crawling by looping
    through the To Crawl list. */
    while (crawling && toCrawlList.size() > 0)
    /* Check to see if the max URL count has
    been reached, if it was specified.*/
    if (maxUrls != -1) {
    if (crawledList.size() == maxUrls) {
    break;
    // Get URL at bottom of the list.
    String url = (String) toCrawlList.iterator().next();
    // Remove URL from the To Crawl list.
    toCrawlList.remove(url);
    // Convert string url to URL object.
    URL verifiedUrl = verifyUrl(url);
    // Skip URL if robots are not allowed to access it.
    if (!isRobotAllowed(verifiedUrl)) {
    continue;
    // Update crawling stats.
    updateStats(url, crawledList.size(), toCrawlList.size(),
    maxUrls);
    // Add page to the crawled list.
    crawledList.add(url);
    // Download the page at the given URL.
    String pageContents = downloadPage(verifiedUrl);
    /* If the page was downloaded successfully, retrieve all its
    links and then see if it contains the search string. */
    if (pageContents != null && pageContents.length() > 0)
    // Retrieve list of valid links from page.
    ArrayList links =
    retrieveLinks(verifiedUrl, pageContents, crawledList,
    limitHost);
    // Add links to the To Crawl list.
    toCrawlList.addAll(links);
    /* Check if search string is present in
    page, and if so, record a match. */
    if (searchStringMatches(pageContents, searchString,
    caseSensitive))
    addMatch(url);
    // Update crawling stats.
    updateStats(url, crawledList.size(), toCrawlList.size(),
    maxUrls);
    // Run the Search Crawler.
    public static void main(String[] args) {
    SearchCrawler crawler = new SearchCrawler();
    crawler.show();
    }                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

  • Sharepoint crawl db growing exceptionally after full crawl run..

    while we run the crawl sharepoint crawl db growing exceptionally.....drive is running out of space.
    lots of errors coming in crawl.....it is trying to do the crawl even on thoese sites which were moved, deleted or URL changed.
    everything is showing in crawl.....any script/suggestion please?

    how many documents you are crawling? full crawl will crawl all sites if its deleted then it delete from index but 1st check and make sure.
    check these post:
    http://social.technet.microsoft.com/Forums/sharepoint/en-US/f0d95c9c-5797-4818-9189-609a48af59d0/crawl-database-size-is-growing-rapidly?forum=sharepointgeneralprevious
    http://www.itwalkthru.com/2011/11/sharepoint-2010-application-crawl-store.html
    for deletion of item from index may be need to play with deletion policy:
    http://technet.microsoft.com/en-us/library/hh127009.aspx
    Please remember to mark your question as answered &Vote helpful,if this solves/helps your problem. ****************************************************************************************** Thanks -WS MCITP(SharePoint 2010, 2013) Blog: http://wscheema.com/blog

  • MacBook Pro Slowed to a CRAWL after less then a week....

    Hello All,
    I'm new to Mac. I just got a MacBook Pro last week.
    Well, for some reason it has slowed to a crawl....
    The dock show/hide and magnify isn't working either. Neither is the "active screen corners" feature on the dashboard. In addition to that, the system does not highlight menu items and show the submenus.
    Also when starting applications they take forever to load.
    It's acting like it has no ram even though it has 2 gigs and a 2.16 processor. The ram is showing up in the "About this MAC" screen though.
    I've tried turning off all startup login items. Nothing is being loaded into memory. I've also tried running the disk utility.... still having the same problems.
    I'm new to Mac so perhaps I'm missing something completely obvious.
    What would you all suggest is going on, and what do I need to do to fix it?
    Thank you
    MacBook Pro   Mac OS X (10.4.6)  

    James:
    I suspect Wizard's problems resulted from too many widgets and possibly some not quite compatible software. Slowdown can result from lack of free physical RAM, so that Virtual Memory swap files on disk have to be frequently accessed. I don't think he mentioned how much RAM is installed. His symptoms go beyond mere slowdown, however, so I would suspect other software problems as well.
    Cutting way back on widgets together with restarting can work wonders sometimes.
    OS X handles file fragmentation pretty well, for file sizes up to about 20 MB. It's probably not necessary to run defrag on a Mac unless one wants to clean up a large space for video work. I don't have defrag software. I'd rather wipe the drive (maybe once a year) and reload from a backup drive. I do swear by DiskWarrior, however, and hope they release an Intel version soon. DW does a great job of correcting disk and file errors and optimizes the directory.
    Apple has been packing a lot of changes into OS incremental upgrades. To interpret that as comparable to Windows Service Packs probably isn't very accurate. Yes, some bugs are fixed but there have been steady upgrades of core features. Security fixes are often included as well. But I've experienced no stability problems whatever.
    Some users recommend using Combo updates instead of using Software Update. I've always used Software Update and have never had a problem. The important thing, either way, is to make certain the operating system and disk directory are 'clean' and trouble-free before and after the update.
    I've never had to do an OS reinstall except once, when I was tinkering with System files and slipped up. Routine preventive maintenance works.

  • Crawl Problems in H.264

    Greetings.
    I just exported a 5-minute sequence from HDV 1080i to H.264 (not with Compressor). The .mov file when played back has problems. The images sometime move clumsily across the screen. The crawl portions, which contain English text in Helvetica font size 23, move at times erratically from right to left. The scrolling credits at the end also appear to jump or flicker at times.
    Are there pre-sets for crawl or scroll text: special fonts, speed, use of Motion or output modes?
    Any suggestions are appreciated.
    Tom

    Are you streaming it from a external HDD?

  • Macbook Pro running slow/(crawling) after 10.9.4 update

    I paste in information on the system from EtreCheck below.
    The Macbook is really crawling very slowly.
    It may be a coincidence, but happened when I restarted after the latest software update 1st july from OSX 10.9.3 to OSX 10.9.4
    Regarding How does it behave/How slow:?
    Now, every task that may require the use of resources not recently used (menus, lists etc.) will take exceedingly long time. For example after the desktop appears with the top menu, it will take 3-4 min before any menus can be pulled down, same with dock, 3-4 min to show up. Clicking an app, 3-4 min before launches. Then after launched, some apps may run OK, for example Logic Pro.
    When working in the Finder. Opening a folder may take 3-4 min, before the the file list shows. Then selecting a file - takes a minute before it gets hi-lighted. But is then fast to drag to an open folder on the backup drive. Copying a folder to a backup drive, is also fast.
    Another example: It took two hours from logging in at this Forum, until getting here to post.!! (I'll read answers from another machine)
    What I have done:
    I have backed up with Time Machine,, in addition to manually backing up files.
    I have ran Disk Utility and it reports the hard  disk is OK. I have ran fix permission, only two were fixed. No difference.
    I have reset the System Management Controller (SMC). No difference.
    It is not a network problem, My network is fast and smooth with other devices,
    I don't want to change hard drive 'just in case'. It fast to restore files from the backup. However
    I  have many licences apps and contents which will take several weeks to restore.  I've done that before.
    Is it likely that it is the hard disk that is failing without disk utility having a clue?
    SYSTEM -INFO:
    EtreCheck version: 1.9.12 (48)
    Report generated 10 Jul 2014 12:06:30 GMT+2
    Hardware Information:
      MacBook Pro (13-inch, Late 2011) (Verified)
      MacBook Pro - model: MacBookPro8,1
      1 2.8 GHz Intel Core i7 CPU: 2 cores
      4 GB RAM
    Video Information:
      Intel HD Graphics 3000 - VRAM: 384 MB
      Color LCD 1280 x 800
    System Software:
      OS X 10.9.4 (13E28) - Uptime: 0 days 2:7:15
    Disk Information:
      TOSHIBA MK7559GSXF disk0 : (750,16 GB)
      EFI (disk0s1) <not mounted>: 209,7 MB
      Macintosh HD (disk0s2) / [Startup]: 749,3 GB (571,85 GB free)
      Recovery HD (disk0s3) <not mounted>: 650 MB
      MATSHITADVD-R   UJ-8A8
    USB Information:
      Seagate Expansion Desk 2 TB
      EFI (disk1s1) <not mounted>: 209,7 MB
      Seagate Expansion Drive (disk1s2) /Volumes/Seagate Expansion Drive: 2 TB (1,74 TB free)
      Apple Computer, Inc. IR Receiver
      Apple Inc. FaceTime HD Camera (Built-in)
      Apple Inc. BRCM2070 Hub
      Apple Inc. Bluetooth USB Host Controller
      Apple Inc. Apple Internal Keyboard / Trackpad
    Thunderbolt Information:
      Apple Inc. thunderbolt_bus
    Gatekeeper:
      Mac App Store and identified developers
    Kernel Extensions:
      [not loaded] com.zoom.R16USBAudioDriver (1.4.0 - SDK 10.6) Support
      [not loaded] jp.co.roland.RDUSB0127Dev (1.0.0) Support
    Startup Items:
      RDUSB0127Startup: Path: /Library/StartupItems/RDUSB0127Startup
    Problem System Launch Daemons:
      [failed] com.apple.msrpc.netlogon.plist
    Launch Daemons:
      [loaded] com.adobe.fpsaud.plist Support
    Launch Agents:
      [invalid] entries
      [running] com.zoom.R16DeviceManager.plist.svn-base Support
      [running] com.zoom.R16DeviceManager.plist Support
    User Launch Agents:
      [failed] com.fastestyoutubedownloader.TrialExpired.GoldFreeFirstWeek.plist Support
    User Login Items:
      iTunesHelper
    Internet Plug-ins:
      FlashPlayer-10.6: Version: 13.0.0.214 - SDK 10.6 Support
      Flash Player: Version: 13.0.0.214 - SDK 10.6 Outdated! Update
      QuickTime Plugin: Version: 7.7.3
      Default Browser: Version: 537 - SDK 10.9
    Audio Plug-ins:
      BluetoothAudioPlugIn: Version: 1.0 - SDK 10.9
      AirPlay: Version: 2.0 - SDK 10.9
      AppleAVBAudio: Version: 203.2 - SDK 10.9
      iSightAudio: Version: 7.7.3 - SDK 10.9
    iTunes Plug-ins:
      Quartz Composer Visualizer: Version: 1.4 - SDK 10.9
    3rd Party Preference Panes:
      Flash Player  Support
      GR-55  Support
      ZOOM R16ControlPanel  Support
    Time Machine:
      Auto backup: YES
      Volumes being backed up:
      Macintosh HD: Disk size: 697.84 GB Disk used: 165.26 GB
      Destinations:
      Seagate Expansion Drive [Local] (Last used)
      Total size: 2
      Total number of backups: 17
      Oldest backup: 2014-07-04 07:43:35 +0000
      Last backup: 2014-07-10 10:39:24 +0000
      Size of backup disk: Adequate
      Backup size 2  > (Disk used 165.26 GB X 3)
      Time Machine details may not be accurate.
      All volumes being backed up may not be listed.
    Top Processes by CPU:
          0% hidd
          0% WindowServer
          0% RDUSB0127Setupd
          0% SystemUIServer
          0% Finder
    Top Processes by Memory:
      219 MB mds_stores
      70 MB Finder
      61 MB Safari
      57 MB WindowServer
      41 MB mds
    Virtual Memory Information:
      1.71 GB Free RAM
      1.15 GB Active RAM
      131 MB Inactive RAM
      1.01 GB Wired RAM
      255 MB Page-ins
      0 B Page-outs

    Thanks Melophage
    You ask some  (5) questions regarding the output of Etrecheck
    Re #4 :"Have you tried uninstalling the fastestyoutubedownloader.com user launch agent, in case that’s causing a problem?"
    Yes, now I have.  Now it takes 40 min. from powerup to the login screen.!
    My system report  showed as I posted, before the removal:
    "User Launch Agents:
      [failed] com.fastestyoutubedownloader.TrialExpired.GoldFreeFirstWeek.plist"
    I am ignorant about the matter.  I too was surprised to find that item and the conspiracy deamon on my left shoulder had wild thoughts.
    The app that generated the Launch Agent must have been FastestYouTubeDownloader.app,
    It was downloaded as an app and no uninstaller. I think it was supposed to run fast  the first week, and if you paid for it(upgrade license), it would remain fast or else work slowly. I tried it on one file and then threw it in the trash, as I would have no more use for it. But a nagging/promo dialog telling me the trial period was over came up 1st july, while the nagging came up I was notified (by appstore) that a new update of my system required a restart. I pressed that, my system was updated to 10.9.4 and after the update everything was slow. So the conspiracy deamon was whispering that maybe the app had made everything slow, by mistake, and not just the FastestYouTubeDownloader.
    I really don't think so.
    But I think the developers have to leave a mark somewhere to tell them that I have tried the software and maybe they have chosen the User Launch Agent for this?  Then trying to remove it may just cause more problems, since it is not anything I should be able to do for obvious reasons.
    I couldnt see what it was really doing from its content, it seems to me it should do nothing if the application was not found.
    Re # 1 and #2
    I have had these drivers all the time and not experienced problems with earlier versions of Mavericks.
    Also, the ZOOM and the GR-55 is not connected.
    Both had at least their own kernel-extension, a launch agent and a preference panel, but possibly more.
    But I have now tried the Uninstaller that accompanies the GR-55 driver  to remove it.  It gave no improvements.
    Re # 5
    I have apperently two Flash Players, both have the same version number, one is outdated.?
    Probably not good, but probably not my main problem(s).? I'm not using Flash often.
    Re # 3
    I have no idea, what the following listing
    Launch Agents:
      [invalid] entries
    may refer to. In my HD/Library/LaunchAgents folder I only have
    "com.zoom.R16DeviceManager.plist" the other entry in the report with same file name with added extension  ".svn-base" I have no idea where is coming from.
    I think this is not my primary problem though? even if  I am curious where it comes from.
    My Conclusions
    Obviously with a boot time of more than 45 min  doing diagosics and tests takes long time.
    The Disk Utility verifies the disk as OK.   Onyx verifies the disk as OK (mabe the same test?).
    I think my maachine is just getting worse.  There must be a serious problem, not  just a fine tuning issue.
    The conspiracy deamon on my other sholder tells me that Apple have decided its time for me to buy a new computer again :-) ,It  would have been a good idea if not for the pain of  reinstalling licensed software which will require a 2 week vacaition doing nothing but restoring, Possibly having to purchase new licences costing more than the computer.

  • How can I get Print to PDF to accept oversized page sizes?

    To level set, my platform is Mac OSX 10.6,  Snow Leopard. 
    I sometimes create banners and such that are designed to be plotted/printed on large format printers.  For example, a banner might print to a page that is 12" x 72".
    Now, I don't actually own such a printer, the FedEx Kinko's down the street does, hence I don't have a driver installed for it. 
    I really don't know the "best" tool for creating these drawings. I have tried MS PowerPoint (which can't seem to accept paper wider than 56 inches) and have  used NeoOffice Draw, which  seems to work ok until it comes to printing time.
    When I try to print to PDF, which is what the folks at Kinko's want, the printer systems always asks if I want to rescale the print to fit my 8-1/2 X 11 home printer(s).  No, just print it "as-is" is NOT a choice and neither is "Any Printer".   I end up having to crawl through setting up odd sized paper under "Any Printer" and somehow, I still don't really understand how, the print to PDF now saves the file sized to print as drawn.
    So, I can see a few possible approaches to making this easier. 
    Install the print driver for the printer at Kinko's.  Only, I don't think the installation process will allow that unless the printer is connected.
    Develop a dongle that fools the Mac into thinking the printer is attached.  Cute idea that could make money if I sold them, but hardly reasonable.
    Develop a dummy (virtual?) print driver that can be added without a physical printer having to be present and which opens a window when you "print" to it that allows you to specify the page size and other information with which it will "lie" to the OS and say, "Yes,  I am a printer capable of doing this....".  (FYI, I am a programmer going on 30 years now.)
    So, before I teach myself Objective-C and how to use Xcode, is there a better or known solution?  Is there an existing product I can use?  Is there a driver development kit for virtual printers?  Any chance I could get a C++ kit and not have to learn Objective-C?
    Thanks!

    Macropanther wrote:
    I don't think the installation process will allow that unless the printer is connected.
    Incorrect. Win or Mac, you can install a printer driver w/o the printer being connected.
    I have tried MS PowerPoint (which can't seem to accept paper wider than 56 inches) and have  used NeoOffice Draw
    You're using the wrong tools. PP shouldn't even be in consideration; I don't know about NeoOffice, but I tried its cousin LibreOffice, and it has issues with custom paper sizes; which is not surprising, given what these apps were designed for.
    What you need for actual banners (not web page banners) is a vector drawing tool or a DTP app. I'd say Illustrator or InDesign, but Adobe CS is expensive, and, if you're thinking of PowerPoint for this job, you're not ready for it.
    You can try Pages; although not quite DTP, it can do the job and it's inexpensive. There are other options, such as Swift Publisher (a DTP app) or Intaglio (a drawing app) -- both are more expensive than Pages, but still affordable; or you can go open source with Scribus (DTP) or InkScape (vector; requires X11). You already know the ups and downs of using open source. (These are just examples; you can find others if you look for them.)
    Although it's better to have Kinko's printer driver, to make sure of the correct options, you don't actually need it to print to PDF. You should be able to define a custom page size (eg, 12in × 72in) and create the PDF from it. Any app which uses the Mac OS X standard print function should be able to do it.

  • Photoshop CS2 slowed to a crawl.

    I'm using CS2 on a G4 Dual 500 with 1gb RAM. I'm running OSX 10.4.11 & Photoshop is my main tool so I have many years experience using it.
    I'm currenly working on a file that's about 500mb & the "scratch sizes" box says 2.12gb.
    As far as I'm concerned this is a fairly familiar situation that I've encountered many times before.
    However, today my machine has slowed to a crawl. It's taking about 15 seconds PER BRUSH STROKE to clone around an image & I can't figure out why.
    I recently bought a 500gb SATA hard drive with a PCI controller card & installed a fresh system on it. (I did however use the option to copy the original hard drive over to the new drive so it wasn't a completely fresh install. I've also recently trashed the CS1 apps that CS2 left sitting on my hard drive but I couldn't see how to do a proper uninstall so I just trashed the apps. from the apps. folder.)
    I don't know if this has anything to do with it but I'm at my wits end, staring a tight deadline in the face with a seemingly inexplicable problem that's come out of nowhere.
    Any ideas?

    Del,
    >I'm using CS2 on a
    G4 Dual 500 with
    1gb RAM
    >I deal with files in
    excess of 3GB on a regular basis without any problem at all.
    >the file was in 16 bit colour not 8 bit
    While I'm not doubting your statements, the numbers in boldface are certainly not working in your favor. 16 bit files only compound the issue.
    I would be surprised if your computer did
    not choke under these circumstances.
    Neil

  • When I full crawl my MOSS 2007 site, it is Intializing my site repeatedly!

    Hi everyone
    I'm using Moss 2007. When I did full crawl, it put out lots of logs. I read one of the logs and found that several strange parts on them.
    It sounds like ...
    1. It is initializing security for my subsite repeatedly.
    2. Cache is full
    3. My FileSharedLibrary allows no one to read items.
    1. It is initializing security for my subsite repeatedly.
    I don't know why it is initializing my subsite again and again. I believe this made the logs so large...
    Here is the log.
    Init security for site "MyItem'sURL" - File:d:\office\source\search\search\gather\protocols\sts3\sts3util.cxx Line:842
    Init Group GROUPNAME123 Id 7 - File:d:\office\source\search\search\gather\protocols\sts3\sts3util.cxx Line:3086 
    New User USERNAME123 Id 1 - File:d:\office\source\search\search\gather\protocols\sts3\sts3util.cxx Line:1073
    New User USERNAME456 Id 13 - File:d:\office\source\search\search\gather\protocols\sts3\sts3util.cxx Line:1073 
    2. Cache is full
    Also I think the log says the cache is full. But I have no idea this cache problem and the initializing problem are effecting each other or not.
    Remove V3 Cache, reason = oversize, committed 649506816 - File:d:\office\source\search\search\gather\protocols\sts3\sts3util.cxx Line:127 
    Remove V3 cache - File:d:\office\source\search\search\gather\protocols\sts3\sts3util.cxx Line:76 
    V3 Cache statistics: committed=649506816, UserCatalogCaches=1 - File:d:\office\source\search\search\gather\protocols\sts3\sts3util.hxx Line:172 
    3. My FileSharedLibrary access control allows no one to read items.
    Thirdly, I got this error many times. (FileSharedLibrary is the feature of DocAve)
    I searched "hr=80041211" and found MicroSoft said "Item will not be indexed. Its access control list allows no one to read the item. Used in GetSecurityDescriptor method of the IUrlAccessor interface.
    CSTS3Accessor::GetItemSecurityDescriptor: Return error to caller, hr=80041211
    - File:d:\office\source\search\search\gather\protocols\sts3\sts3acc.cxx Line:1210 
    I already checked DisabledLoopBackCheck is disabled, Crawl Account have a read access of my site properly, and I'm sure we have enough free space at C and D drives.
    Actually I can use my SharePoint site normally with no problem and it does not show any error on my site. But I really need to know the reason why this three points are happening...in order to avoid a huge-size log gives damage to server when I full
    crawl next time...
    Hope someone could help me to solve this problems

    Thank you Jaggi
    I think this is the almost same page that i checked already and mensioned in my question
    "hr=80041211" and found MicroSoft said "Item will not be indexed. Its access control list allows no one to read the item. Used in GetSecurityDescriptor method of the IUrlAccessor interface.
    Do you think my problem is really because of anything Access control problem..?

  • HUGE file size in InDesign CS5

    Help! Deadline fast approaching and InDesign now down to a crawl. It's definitely InDesign... it's open while I'm typing this and I can see the letters here appear as I type. In InDesign, it's like being back in the dark ages... type way ahead and wait for the display to catch up, and it doesn't even get all the way. I've resorted to typing in Notepad and copy/pasting in to try to speed things up a bit.
    I'm a relatively new InDesign user, so apologies if there is a simple answer...
    My file was getting large in terms of number of pages and because of the anchored frames issue (another post), I split it into several files and put it in a book. At least that way the knock on effect of having something NOT anchored was manageable.
    Yesterday and now today, the program got progressively slower. Using 70%+ of memory (4GB of it!)
    Something feels very wrong to me...
    The first file in the book is 4 pages long. It consists of a title page, an automatically generated ToC, a whole-page frame with a border and a whole-page frame without a border. Apart from the title page, it's ALL text. On the title page and the master page I have (linked, I believe - they show up in the links pane) two images. The one on the front cover is only 115KB anyway, and the one on the master page, in the footer, is 82KB.
    This file is  13,776KB in size. What's going on? The whole BOOK before was slightly less than 12MB, and although I've added a bit more content, nothing new has gone into these four pages. The total for the whole book now (NOT INCLUDING IMAGES) is just over 35MB.
    I had installed a trial of CS 5.5 to see if that was going to be any good for an ePub project I'm about to start, but until recently I hadn't noticed any problems with it. I've just uninstalled it and am about to do a repair of InDesign CS5 to see if that helps.
    For the technically minded:
    PC running Windows 7 Professional 32-bit,
    Intel i5 CPU (3.84GHz)
    Two graphics cards running three monitors (been like this since got the PC last December).
    I'm going to struggle to meet the deadline at this rate, and am starting to panic :-(
    Thanks in advance for any help/suggestions

    You're a star. I'd already done the switching off the page thumbnails and reducing page performance. I hadn't thought about cross-references! That's certainly a possibility.
    I've just done the "Save As" twice - and that first file is down to 1,100 KB - which is a huge reduction from it's previous size! It's still growing already though - just updating the numbers in the book has taken it from 800 KB up to the new 1,100 without me doing anything else (not even opening it myself). Track Changes is (I believe) switched off ... in Edit > Preferences, unless there's somewhere else...
    Shoudl have thought of the Save As as well - used to have to do that with Word docs a few versions ago (not so much now).
    Thanks again.
    Alison

  • Crawl Log Error: "The filename or extension is too long."

    Hello All,
    I get this error for some files on my SharePoint 2010 error log. The fix suggested
    here does not work as its for earlier versions of sharepoint 2010. I have verified the files in question have a title lenght of 120 characters and Name(Link) with character lenght of 71.
    Whats weird is sharepoint indexes some similarly named files but exempts the other. These files are basically the same just named different as they belong to different users.
    Any help?

    It could be because of this also:
    Assume that a custom list in a Microsoft SharePoint Foundation 2010 site or in a Microsoft SharePoint Server 2010 site contains a multiline text column. If the data size of the custom column is more than 100 kilobytes (KB), only the first 100 KB of the data
    is indexed when you perform a crawl on the SharePoint site. Additionally, you receive the following error message:
    The filename or extension is too long.
    RESOLUTION : http://kbalertz.com/2288793/filename-extension-error-message-crawl-perform-crawl-SharePoint-Foundation-SharePoint-Server.aspx

  • .pdf file size limits

    Is there a limit to .pdf file size or page count?

    Just to confirm what Bill said —
    I have several 'boxes' that I use each day.
    They range from the typical, adequate, lower price point to a high end 'power-box'.
    I have occasion to combine a large PDF document collection into a single PDF having a foot print of just under 800Mb (page count out around 2,000).
    When using the 'typical, adequate, lower price point' box this large PDF can only be consumed at a crawl.
    Bottom line is that the typical box has less RAM, resources, and integrated graphics rather than a adequately robust independent graphics card.
    These boxes just do not provide the 'logistical' support to applications (Adobe Reader, Acrobat, whatever) that render PDFs with the very high file size foot print.
    Particularly so if the PDF contains a good dose of graphics.
    For the most part I've found that end-users seriously avoid such PDFs.
    Be well...
    Message was edited by: CtDave

  • "Wrong Size" in and Unavilable Pulldowns in LR  Prints in PS

    I have successfully printed in LR before but this time I'm running into weird problems. I tried printing an A3 size print (11.7 x 16.5) Landscape in LR, set the paper size to A3 (Retain Size, Maximum). I noticed that the cell size in LR read 11.92 for the Height even though the paper itself is only 11.7.
    I tried printing anyway, my Epson Stylus 4800 Pro started its printing routine but then stopped and gave me a dialog box saying the paper size was wrong. I changed the cell Height to 11.5 and the same thing happened. Refuses to print because paper size is wrong, which it says on my Epson's screen.
    Also, when I was setting up the job and pulled down the menu to ensure that Color Management was off, that choice was lined-out along with Paper Config, Roll Paper Option, and Expansion. When I selected one of those, a dialog popped up saying that Seiko/Epson didn't include these options in the software. I know that's not true because I've used them before.
    So I sent the file to Edit in PS, sized the photo, and printed it out with no problem. The items which were lined-out were not lined-out in PS.
    I'm got Mac Pro, 10.5.6 OS, and my Epson Stylus Pro 4800. And as I said, I've successfully printed other sizes before in LR.
    Has anyone had this happen to them and/or do you have any idea what is going on?

    Your suggestion of switching to 32 bit solved the pulldown items being lined-out. Also now can print without the dialog box telling me that the size is too large.
    However, now that I've added a number of changes with the Adjustment Brush, LR has slowed down to an unacceptably slow crawl. When I've tried using the Graduated Filter tool, LR is so slow that when I try to move or rotate the gradient I can't reliably see the adjustment for so long that I can't reliably use it.
    I'm using a Mac Pro with 2 x 3.2 GHz Quad- Core Intel Xeon, 16 GB 800 MHz DDR2 FB-DIMM. I've read postings here regarding similar slowness issues with LR, but is what I'm experiencing "normal" for this release of LR? If so, it is a sad commentary on the quality of the product.

  • Crawl perfect in FCE4, jittery on export

    Hi-
    I'm creating crawls over Hi Res photos in FCE4 that are smooth in my Canvas but when I export them to .mov there's a slight jitter.  I've tried just about every export parameter and still no change.  I've also tried viewing the export in Quicktime Player, VLC and even tried burning to a DVD with Toast.  All exhibit the same problem.
    Thanks for the help!
    Some Specs:
    Mac Mini 10.5.8
    Processor: 1.83 GHz Intel Core 2 Duo
    Memory: 3GB 667 MHz DDR2 SDRAM
    Creating project in 1440 x 1080
    Creating crawl in Live Type, then tried Boris plug in

    The jitter in the lettering is due to the interlaced nature of the video.  It generally does not appear in the Canvas but does in the finished video.
    There's not much you can do about it in FCE other than use a larger font size, and even that is not going to completely eliminate it.
    In Boris Title Crawl there is a control called 1:2:1 Deflicker - try checking it and it may help decrease the amount of jitter, but it will not eliminate it.

  • Help Required on Crawling Database

    Hi Experts,
    I am working in a Web-based Environment on Windows 2003 R2 and Oracle 9.1. For the past some time, I have been experiencing a crawling Oracle database. The reports which used to take a few seconds are now taking minutes. The users are not able to connect to the database and complain of very slow database. Since, I am new Database Administration, Can any of the expert help me in tuning the database for maximum throughput. For assistance, I am providing with the Health Check of my Database extracted through TOAD.
    *** BEQ-LOCAL                             07/09/2011 6:31:50 PM  ***
    ~Time to Connect             : 0 seconds.
    ~Database Version            : Oracle9i Enterprise Edition Release 9.0.1.1.1
    ~Database Up Since           : 01:01:09 PM, September 7 2011
    !Buffer Cache Hit Ratio      : 77.8311   <------- (in red colour)
    ~Library Cache Miss Ratio    : 0.2612
    ~Dictionary Cache Miss Ratio : 2.7429
    [Shared Pool Usage] Exec Time 0 seconds
    ~ Total Mb Unused : 23.97
    ~ Total Mb Used   : 24.03
    ~ Total Mb        : 48
    ~ Shared Pool Percent Used : 50.06
    [Archive Log Mode Info] Exec Time 0 seconds
    ! Archiver : STOPPED   <------- (in red colour)
    ! Log Mode : NOARCHIVELOG <------- (in red colour)
    ! log_archive_start (init.ora param) = FALSE <------- (in red colour)
    [Archive Log Info] Exec Time 0 seconds
    ~ Average Log Switches Per Day                          : 12.58
    ~ Hard Drive Storage (in Mb) for this many archive logs : 1257.92
    !  Error using UTL_FILE to examine alert log! <------- (in red colour)
    !  Error using UTL_FILE  <------- (in red colour)
    !  Possible causes:  <------- (in red colour)
    !  1)  You don't have privileges to execute the UTL_FILE package.  <------- (in red colour)
    !  2)  UTL_FILE_DIR initialization parameter does not include one of these lines:   <------- (in red colour)
    !      utl_file_dir=D:\oracle\admin\orcl90\bdump   <------- (in red colour)
    !      utl_file_dir=*    <------- (in red colour)
    !  3)  alert.log file name is not among the following:    <------- (in red colour)
    !      (if this is the case please inform Quest support)   <------- (in red colour)
    !      alert_orcl90.log  <------- (in red colour)
    !      orcl90alrt.log   <------- (in red colour)
    !  4)  No directory has been created for D:\oracle\admin\orcl90\bdump,   <------- (in red colour)
    !      or directory exists but privileges have not been granted for it.    <------- (in red colour)
    [Redo Log Group Sizes and Quantities] Exec Time 0 seconds
    ~ Number of Log Groups : 3
    ! Number of Members per Log Group: 1    <------- (in red colour)
    ~ All redo log members are 100M in size.
    [Rollback Segments with wait ratios > 1 %] Exec Time 1 seconds
    ~ None
    [Objects with Mixed-Case Names] Exec Time 1 seconds
    ! Synonym            PUBLIC.NameFromLastDDL   <------- (in red colour)
    [Tables with > 5 % chained rows and > 500 total rows] Exec Time 0 seconds
    ~ None
    [Table Partitions with > 5 % chained rows and > 500 total rows] Exec Time 0 seconds
    ~ None
    [Segments with < 10% of extents remaining and (maxextents > 1)] Exec Time 11 seconds
    ~ None
    [Segments with > 100 extents] Exec Time 4 seconds
    ~ None
    [Objects which can't extend because there is not
    enough room in the tablespace] Exec Time 3 seconds
    ~ None
    [Jobs] Exec Time 0 seconds
    ~ None
    [redundant object privs with conflicting grant option] Exec Time 2 seconds
    ~ None
    [Profiles that are not granted to any user] Exec Time 0 seconds
    ~ None
    [Tablespace Fragmentation] Exec Time 0 seconds
    ~ None
    [Tablespaces with less than 10% free space] Exec Time 1 seconds
    ! EXAMPLE                   : 0.1% Free Space,  MB Free / Total : 0 / 153    <------- (in red colour)Thanks

    Install statspack and create a performance report and post the output here
    Installing and Configuring StatsPack Package          [Document 149113.1]
    Creating a StatsPack performance report          [Document 149124.1]
    FAQ- Statspack Complete Reference          [Document 94224.1]
    Systemwide Tuning using STATSPACK Reports          [Document 228913.1]
    HTH
    Srini

Maybe you are looking for