Flat file data sources info pack executing for long time with yellow

I can see flat file all records 140 in above screen but i could not see data in PSA . when i am trying to see the PSA it show not data .
I have tries to make status to green but still request not process to Data target .

HI,
When you see message like there is no psa mean data source was not in active.
so activate your data source Thru by using SE38 - RSDS_DATASOURCE_ACTIVATE_ALL(bw 7.x).
After the activation you can try to load data.before that make your yellow request to red status and delete it.
later reload the same flat file.
Thanks

Similar Messages

  • Urgent: Transport of Flat File Data Source

    While transporting flat file data source is it required to transport first in a separate request or everything can go in 1 request.
    Thank you,
    sam

    Hi,
    In this case it can go in one request as no replication is needed. But for other datasources, you need to send the datasources first in a separate request and then replicate the datasource in BW before you send the BW changes.
    Cheers,
    Kedar

  • Flat file data Source creation in BI7

    Hi ,
    I am trying to create flat file data source in BI7 to load master data.
    Aftre i enter all the details,while activating its giving me error like this
    Runtime Errors         DATA_LENGTH_NEGATIVE
    Except.                CX_SY_RANGE_OUT_OF_BOUNDS
    Can anyone suggest me abt this.
    Thanks
    Thiru

    Hi,
    In the datasource def go to the fields tab. Here check the values for the length og the field.
    Check whether all the values are within 60
    Regds,
    Shashank

  • Workflow trf executing for longer time

    Hi All,
    Workflow trfc executing for longer time in transaction SM58. Why will this happen ? Is there any resolution available ? Please advise on this issue. Below are the details :
    Caller: WF-BATCH
    Function Module: SWW_WI_EXECUTE_INTERNAL_RFC
    Status Text: Transaction Executing
    Program: SAPMSSY1
    Kind Regards
    Shiva

    Hi,
    Check if WF-BATCH user is not locked or the pwd changed. The same as the process started, see the event trace if you are triggering any wflow from this. And CLick on the event and see the "REceiver Data" and "Trace Data" if any errors are there.
    Sometimes, if this user got locked for some reason, it results in issues like you mentioned.
    Hope this helps
    venu

  • DISASSEMBLE is being executed for long time in SPAM

    Hi ,
    I am trying to applying SAP_BASIS patch in new XI 3.0 server .But it got hanged in <b>DISASSEMBLE is being executed</b> for long time in SPAM.
    Can any one will me to solve in this issue.
    Thank You !
    Gopnath.G

    Hi
    The following Patch manager may be useful
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/BCUPGOCSSPAM/BCUPGOCSSPAM.pdf
    Regards,
    Sai

  • Cube like aggregation behavior from a flat file data source

    Is the following workflow possible in Xcelsius and if so how would your organize you logic to be able to acheive this?  Any examples would greatly be appreciated
    My data source is XML, therefore a flat file.
    The flat file contains multiple dimensions for which the customer wants to aggregate by based on the selection and also include and "ALL" option for every dimension/selector.
    Essentially, the file is structured like this:
    ie: Dim1  Dim 2  Dim3   Dim 4   Dim5  Metric1    Metric2
           a        c         e          j          l          5             5
           a        c         e          j          m        3             2
           a        c         e          j          n         4             8
           a        c         e         k          o         3             4
           a        c         e         k          n         8             2
            ...etc         
    We need drop down selectors for Dim1-Dim4 with the "ALL" option.  I would have my tables and charts etc use the Dim5, Metric1 and Metric2. 
    I've used the "Lookup" function to create 1 level of aggregation in a flat file structure with 1 dimension, such as the "Lookup" would call all the Dim1entries and insert into a group of referenced cells and then i would use the "Sum" function from Excel.
    But is this possible with multiple dimensions and include the "ALL" option.  If so, how?  The workflow defined is ideal for a CUBE architecture but for many different political reasons we need to be able to address this with Xcelsius being sourced from a flat file ie XML.  Any insight would be greatly appreciated..

    hi there.
    if you backend is SAP BW connected via OLAP Universe you could use L0 Dimensions for all aggregation in WebI Layer. (Check the Blog's of Ingo Hilgefort for details)
    Best Regards
    Ulrich
    http://www.xcelsius-insight.com

  • SP Executes For Long Time

    Hi There,
    I'm having 150 lines of SQL statements (SQL Server 2000),
    with 5 parameters, executes in 12 seconds. But when I make a Stored
    Procedure out of it (with exactly the same 150 lines of code) with
    5 input paramentes, the execution take from 5 to 10 minutes.
    There is a huge difference in execution time 12 sec vs 10
    minutes for exectuing the same code. Has anyone experianced this
    before?
    Many thanks

    You can tell Apple directly at the link below.
    http://www.apple.com/feedback/appletv.html

  • Have been using FaceTime for long time with same person, for the last two weeks when trying to connect, we get connection failed message, something change in settings that we should look at?

    Now getting FaceTime failed message when trying to connect with a person that were connecting successfully with up unto two weeks ago, has something changed?

    Take a look around you here in this forum. Many people are talking about having problems using FaceTime. A few users claim to have contacted Apple and Apple says that they are aware of an issue with FaceTime and that they are working on it.
    Some users claim that upgrading to iOS 7 solved there problem with FaceTime. Take a look around the forum and read some of the other discussions.

  • SSIS 2012 - Double quote charcter in data does not allow to load flat file to source component

    Hello everyone,
    I've created a source component for flat file data.
    This file contains data the column data in double quotes. Like,
    First Name Last Name
    "Ankit"       "Shah"
    The text qualifier is double quote for source file connection. 
    When I do have data like below, it does not able to load the source file. And throws an error "The column delimiter for column was not found"
    First Name Last Name
    "Ank"it"       "Shah"
    Can anybody please advise how to resolve this problem?
    Any help would be much appreciated.
    Thanks,
    Ankit
    Thanks, <b>Ankit Shah</b> <hr> Inkey Solutions, India. <hr> Microsoft Certified Business Management Solutions Professionals <hr> http://ankit.inkeysolutions.com

    Please refer:
    http://visakhm.blogspot.com/2014/06/ssis-tips-handling-embedded-text.html
    http://www.ideaexcursion.com/2008/11/12/handling-embedded-text-qualifiers/
    Cheers,
    Vaibhav Chaudhari
    [MCTS],
    [MCP]

  • Error when deploying JDBC file data source for J2EE agent in weblogic

    Hi ODI experts!
    we're using ODI 11g and we've installed J2EE agents on weblogic. We need to configure a connection pool within weblogic for a file data source. Thus, we've added snpsFile.jar (that was shipped with ODI 10g...) to the weblogic config so that we can configure a JDBC connection to file.
    As we validate the config, an error occurs (see below):
    Obviously, driver is found but cannot be deployed.
    ####<4 août 2011 09 h 39 CEST> <Error> <Deployer> <ppdlodi001> <SRVPPDODI_001> <[ACTIVE] ExecuteThread: '2' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <c53bfd6260db74bb:-485a441a:13193b42397:-7ffc-0000000000000027> <1312443551325> <BEA-149265> <Failure occurred in the execution of deployment request with ID '1312443550020' for task 'weblogic.deploy.configChangeTask.6'. Error is: 'weblogic.application.ModuleException: '
    weblogic.application.ModuleException:
    at weblogic.jdbc.module.JDBCModule.prepare(JDBCModule.java:290)
    at weblogic.application.internal.flow.ModuleListenerInvoker.prepare(ModuleListenerInvoker.java:199)
    at weblogic.application.internal.flow.DeploymentCallbackFlow$1.next(DeploymentCallbackFlow.java:508)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:41)
    at weblogic.application.internal.flow.DeploymentCallbackFlow.prepare(DeploymentCallbackFlow.java:149)
    at weblogic.application.internal.flow.DeploymentCallbackFlow.prepare(DeploymentCallbackFlow.java:47)
    at weblogic.application.internal.BaseDeployment$1.next(BaseDeployment.java:1223)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:41)
    at weblogic.application.internal.BaseDeployment.prepare(BaseDeployment.java:367)
    at weblogic.application.internal.SingleModuleDeployment.prepare(SingleModuleDeployment.java:43)
    at weblogic.application.internal.DeploymentStateChecker.prepare(DeploymentStateChecker.java:154)
    at weblogic.deploy.internal.targetserver.AppContainerInvoker.prepare(AppContainerInvoker.java:60)
    at weblogic.deploy.internal.targetserver.operations.ActivateOperation.createAndPrepareContainer(ActivateOperation.java:208)
    at weblogic.deploy.internal.targetserver.operations.ActivateOperation.doPrepare(ActivateOperation.java:98)
    at weblogic.deploy.internal.targetserver.operations.AbstractOperation.prepare(AbstractOperation.java:217)
    at weblogic.deploy.internal.targetserver.DeploymentManager.handleDeploymentPrepare(DeploymentManager.java:749)
    at weblogic.deploy.internal.targetserver.DeploymentManager.prepareDeploymentList(DeploymentManager.java:1216)
    at weblogic.deploy.internal.targetserver.DeploymentManager.handlePrepare(DeploymentManager.java:218)
    at weblogic.deploy.internal.targetserver.DeploymentServiceDispatcher.prepare(DeploymentServiceDispatcher.java:160)
    at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer.doPrepareCallback(DeploymentReceiverCallbackDeliverer.java:171)
    at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer.prepare(DeploymentReceiverCallbackDeliverer.java:41)
    at weblogic.deploy.service.internal.statemachines.targetserver.AwaitingContextUpdateCompletion.callDeploymentReceivers(AwaitingContextUpdateCompletion.java:164)
    at weblogic.deploy.service.internal.statemachines.targetserver.AwaitingContextUpdateCompletion.handleContextUpdateSuccess(AwaitingContextUpdateCompletion.java:66)
    at weblogic.deploy.service.internal.statemachines.targetserver.AwaitingContextUpdateCompletion.contextUpdated(AwaitingContextUpdateCompletion.java:32)
    at weblogic.deploy.service.internal.targetserver.TargetDeploymentService.notifyContextUpdated(TargetDeploymentService.java:225)
    at weblogic.deploy.service.internal.DeploymentService$1.run(DeploymentService.java:190)
    at weblogic.work.SelfTuningWorkManagerImpl$WorkAdapterImpl.run(SelfTuningWorkManagerImpl.java:528)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:201)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:173)
    Caused By: weblogic.common.ResourceException: com.sunopsis.jdbc.driver.file.a.i
    at weblogic.jdbc.common.internal.ConnectionEnvFactory.createResource(ConnectionEnvFactory.java:263)
    at weblogic.common.resourcepool.ResourcePoolImpl.makeResources(ResourcePoolImpl.java:1193)
    at weblogic.common.resourcepool.ResourcePoolImpl.makeResources(ResourcePoolImpl.java:1117)
    at weblogic.common.resourcepool.ResourcePoolImpl.start(ResourcePoolImpl.java:244)
    at weblogic.jdbc.common.internal.ConnectionPool.doStart(ConnectionPool.java:1110)
    at weblogic.jdbc.common.internal.ConnectionPool.start(ConnectionPool.java:147)
    at weblogic.jdbc.common.internal.ConnectionPoolManager.createAndStartPool(ConnectionPoolManager.java:386)
    at weblogic.jdbc.common.internal.ConnectionPoolManager.createAndStartPool(ConnectionPoolManager.java:326)
    at weblogic.jdbc.module.JDBCModule.prepare(JDBCModule.java:254)
    at weblogic.application.internal.flow.ModuleListenerInvoker.prepare(ModuleListenerInvoker.java:199)
    at weblogic.application.internal.flow.DeploymentCallbackFlow$1.next(DeploymentCallbackFlow.java:508)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:41)
    at weblogic.application.internal.flow.DeploymentCallbackFlow.prepare(DeploymentCallbackFlow.java:149)
    at weblogic.application.internal.flow.DeploymentCallbackFlow.prepare(DeploymentCallbackFlow.java:47)
    at weblogic.application.internal.BaseDeployment$1.next(BaseDeployment.java:1223)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:41)
    at weblogic.application.internal.BaseDeployment.prepare(BaseDeployment.java:367)
    at weblogic.application.internal.SingleModuleDeployment.prepare(SingleModuleDeployment.java:43)
    at weblogic.application.internal.DeploymentStateChecker.prepare(DeploymentStateChecker.java:154)
    at weblogic.deploy.internal.targetserver.AppContainerInvoker.prepare(AppContainerInvoker.java:60)
    at weblogic.deploy.internal.targetserver.operations.ActivateOperation.createAndPrepareContainer(ActivateOperation.java:208)
    at weblogic.deploy.internal.targetserver.operations.ActivateOperation.doPrepare(ActivateOperation.java:98)
    at weblogic.deploy.internal.targetserver.operations.AbstractOperation.prepare(AbstractOperation.java:217)
    at weblogic.deploy.internal.targetserver.DeploymentManager.handleDeploymentPrepare(DeploymentManager.java:749)
    at weblogic.deploy.internal.targetserver.DeploymentManager.prepareDeploymentList(DeploymentManager.java:1216)
    at weblogic.deploy.internal.targetserver.DeploymentManager.handlePrepare(DeploymentManager.java:218)
    at weblogic.deploy.internal.targetserver.DeploymentServiceDispatcher.prepare(DeploymentServiceDispatcher.java:160)
    at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer.doPrepareCallback(DeploymentReceiverCallbackDeliverer.java:171)
    at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer.prepare(DeploymentReceiverCallbackDeliverer.java:41)
    at weblogic.deploy.service.internal.statemachines.targetserver.AwaitingContextUpdateCompletion.callDeploymentReceivers(AwaitingContextUpdateCompletion.java:164)
    at weblogic.deploy.service.internal.statemachines.targetserver.AwaitingContextUpdateCompletion.handleContextUpdateSuccess(AwaitingContextUpdateCompletion.java:66)
    at weblogic.deploy.service.internal.statemachines.targetserver.AwaitingContextUpdateCompletion.contextUpdated(AwaitingContextUpdateCompletion.java:32)
    at weblogic.deploy.service.internal.targetserver.TargetDeploymentService.notifyContextUpdated(TargetDeploymentService.java:225)
    at weblogic.deploy.service.internal.DeploymentService$1.run(DeploymentService.java:190)
    at weblogic.work.SelfTuningWorkManagerImpl$WorkAdapterImpl.run(SelfTuningWorkManagerImpl.java:528)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:201)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:173)
    Thanks in advance for your help.
    Fredd.

    Hi!
    Some information: I found this website [http://gerardnico.com/doc/odi/webhelp/en/refmanual/connexion/jdbcdriversample.htm] explaining that the 'old' snpsFile.jar driver was deprecated and that a new one was existing... but I can't find it anywhere!
    My questions are:
    - Can the problem (original problem. See initial post) be due to driver version?
    - Does someone know where to find this 'new' driver?
    Thanks for your help.
    Fredd.
    Edited by: fredd2 on Aug 5, 2011 10:04 AM
    Edited by: fredd2 on Aug 5, 2011 10:07 AM

  • How to specify  inclusion and exclusion rules for File data sources

    This is the seed URL for a file data source: file://localhost/c:/myDir/
    I want to exclude indexing and searching of files under: file://localhost/c:/myDir/obsolete/
    What is the exact format for the exclusion URL?
    I have tried both file://localhost/c:/myDir/obsolete/ and /myDir/obsolete/
    but neither of it seems to work; it still indexes everything under /myDir/
    Should I just put: /obsolete/ as the exclusion URL?
    Also after initial crawling, if I change the inclusion and or exclusion rules and then run the crawler again, it should update the indexes accordingly. Is that right?
    The version of UltraSearch I am using is 1.0.3.
    Thanks for any help on this.

    Try "/c:/myDir/obsolete/"
    Changing inclusion/exclusion rule does not affect files already crawled. It onyl
    affects next crawling behavior.
    To do any DML to existing data set, use SQL directly on wk$url table under the instance owner.

  • Flat file data load - ODS - Look up data in Startroutine

    Hi All,
    There is a requirement that I have two ODS say 1) ABC and XYZ.
    For Both ODS , We load flat file data.
    First we load data to ABC for current fiscal period
    During the data load to XYZ, We lookup part nos data in ABC for current fiscal period , if data there , for those part nos we load XYZ Flat file data.
    My requirement is that , I need to fetch part nos data from ABC for Fiscal period < current fisacl period.  Then I need to extract already loaded data for those part nos from XYZ and Mark the flag field as Non reportable .
    How can we achieve this ?
    Please advice
    Thanks
    Ajay

    Hi ,
    Thanks for your reply.
    I have done so. When I add lines of My Internal table data to DATA_PACKAGE, It gives the syntax error that both structures are not unique. We have several fields in XYZ ods but they are not in Communication structure.
    Thats the problem. Also , if I write the code , will it be executed for each datapackage. I mean , for each datapacket process , my code will fetch whole data from ABC and then from XYZ. Repetion will be there?
    Any other logic can i use ?
    Thanks
    Ajay

  • How to Merge Flat file data and DSO Data

    I have one DSO A which has more than 20 objects (5 of them are keys), which include 0comp_code,0gl_account,.  I have a Flat file which has the following fields ,COMPANY CODE, GLACCOUNT and SUPERVISOR ,EXECUTIVE,ANALYST. My requirement is to get a DSO with consolidated data from DSO A and Flat File data. That means for exapme, In my DSO A , I have US01 COMPANY CODE so when I merge the data in DSO C It should diplay the data for the supervisor exe and analyst for DSO A as well
    When I added with 2 DSOs A (ECC ) and B (Flat file)..C is getting the data from B and its appending the data for Flat file at the end...So I want to merge both...DSO A should match  up with DSO B flat file company code and gl account and should display the remaining fields
    I think its clear..Can any one let me know the solution ASAP...
    Note: We are on BI7.0 not 3.5...
    Thanks in advance

    I have the following key fields in DSO A :
    0COMP_CODE,OGL_ACCOUNT, 0CHRT_ACCT,0FISCVARNT,0FISCPER,0AC_DOC_NUMBER,0ITEM_NUM.
    Here are the total fields in Flat File:
    COMPANY CODE,GL ACCOUNT,SUPERVISOR,ANALYST,EXECUTIVE.
    Here am taking 0COMP_CODE  and 0GL_ACCOUNT  as key fileds in DSO B for Flat File...
    For every month there will be 5 to 6000 records for Flat file ...
    Thanks

  • Flat File as Source in OWB Mapping.

    Hi All,
    Configuration Details is as follows:
    Oracle 9i Warehouse Builder Client: 9.2.0.2.8
    Oracle 9i Warehouse Builder Repository: 9.2.0.2.0
    Windows XP Professional
    I developed a mapping which involves a flat file as source & Oracle Table as target steps followed were as follows.
    1. Created Flat File module, selected Connector & Location Details.
    2. Done mapping for source file & target table.
    3. Validated & Generated the mappings.
    4. Deployed the mapping in Deployment Manager.
    5. Following was Error Message after execution of the mapping.
    Completion Status: Completed with errors.
    Meesage Log: Starting Execution of VHM_MAP2. Starting Task VHM_MAP2. RPE-01013:SQL Loader reported error condition,number 4. Completing Task VHM_MAP2.
    Following was Control File generated by the above mapping.
    OPTIONS ( ERRORS=50, BINDSIZE=50000, ROWS=200, READSIZE=65536)
    LOAD DATA
    CHARACTERSET WE8MSWIN1252
    INFILE '{{TEXT_DATA_LOCATION.RootPath}}dept.txt'
    BADFILE '{{TEXT_DATA_LOCATION.RootPath}}dept.bad'
    DISCARDFILE '{{TEXT_DATA_LOCATION.RootPath}}dept.discard'
    INTO TABLE "{{TARGET_LOCATION.Schema}}"."DEPT"
    APPEND
    REENABLE DISABLED_CONSTRAINTS
    FIELDS
    TERMINATED BY ','
    OPTIONALLY ENCLOSED BY '"'
    "DEPTNO" POSITION (1) INTEGER EXTERNAL ,
    "DNAME" CHAR ,
    "LOC" CHAR
    Please note that no message is displayed in Row Activity & output parameters in Deployment Manager.
    I worked on lower version of the OWB & in that version there was no concept of Location, Connector, Deployment Manager.
    Not sure why the above mapping is failing?
    Also I would like to know where can I get more information about the mapping errors which are different from oracle error messages.
    e.g. In above case error message is RPE-01013: SQL Loader reported error condition, number 4 .
    Thanks in Advance.
    Regards,
    Vidyanand

    Hi
    I'm using single host implementation OWB10g/Oracle on same host.
    My flat file location is
    c:\source\dept.txt
    Pl tell me if I have to save the same file on any particular location for OWB to read.
    Implementation in OWB10g
    1. Created Flat File module, selected Connector & Location Details.
    2. Done mapping for source file & target table.
    3. Validated & Generated the mappings.
    4. Deployed the mapping in Deployment Manager.
    5. Following was Error Message after execution of the mapping.
    Can not read data from file
    c:\souredept.txt
    OPTIONS ( ERRORS=50, BINDSIZE=50000, ROWS=200, READSIZE=65536)
    LOAD DATA
    CHARACTERSET WE8MSWIN1252
    INFILE '{{TEXT_DATA_LOCATION.RootPath}}dept.txt'
    BADFILE '{{TEXT_DATA_LOCATION.RootPath}}dept.bad'
    DISCARDFILE '{{TEXT_DATA_LOCATION.RootPath}}dept.discard'
    INTO TABLE "{{TARGET_LOCATION.Schema}}"."DEPT"
    APPEND
    REENABLE DISABLED_CONSTRAINTS
    FIELDS
    TERMINATED BY ','
    OPTIONALLY ENCLOSED BY '"'
    "DEPTNO" POSITION (1) INTEGER EXTERNAL ,
    "DNAME" CHAR ,
    "LOC" CHAR
    Thx in advance

  • Working with flat file as source in owb 10.2

    Hi,
    I am working with flat file as source . While validating the mapping i am getting the following error like
    " to specify a data file configure the mapping , add a node under ' Source data file', type in the file name
    and select the file location."
    Please give me the suggestion . It is very urgent.

    Hi Venkat,
    I tried the following stepts.
    1. in Design Center select your mapping and right click and select configure
    2. select sql loader data files and select create
    3. On right hand side data file Name : enter your source file name (ex : source.csv)
    4. click ok button.
    5. open mapping and validate.
    The mapping is validating. After validating I deployed the mapping. Up to this the mapping is working fine.
    But when I start the mapping, It has completed with errors.
    The error message is:
    Status
    Error Log
    RPE-01013: SQL Loader reported error condition, number 1.
    LRM-00112: multiple values not allowed for parameter 'control'
    Job Summary
    Updated : 2009-02-24 15:32:43.0 Job Final Status : Completed with errors Job Processed Count : 1 Job Error Count : 1 Job Warning Count : 0
    Please give me the suggestions.
    Thanks,
    Venkat

Maybe you are looking for

  • Help needed with valueChangeListener problem

    I have the following code: <h:dataTable border="0" id="constituyentes" value="#{db_inserta.constituyentes}" var="vl" style="text-align: center;" bgcolor="#F5F5F5" cellpadding="8">      <h:column>           <f:facet name="header">                <h:ou

  • Nice Board for workstations

    Just wanted to post the following. If you are trying to put together a low cost workstation with decent performance. The 890GXM-G65 is a really good choice. On-board video, am3 / am3+ socket, micro-atx with a cost just over $100. Just installed two i

  • Classic scenario CUF to R3

    I've been reading about Extended Classic Scenario (ECS) and how to get customer fields (CUF) to R3 using a BADI. Is something possible for the classic scenario also? I've defined CUF in INCL_EEW_PD_ITEM_CSF* and they show up, but I want to map the CU

  • Cisco Prime Collaboration and SIP Codes

    I am trying to position Cisco Prime Collaboration with a Cisco CUBE router and the client wants to know if Prime will be able to do the following: 1) Scan, index and alert on configurable SIP 4xx, 5xx, 6xx 2) Look at acive inbound/outbound calls

  • Number columns in Pages 5.5

    I am trying to print a table in Pages 5.5 and need to keep the number column that shows up on the left hand side of the page.  Does anyone know how to do this?