RE:naming conversion in real time

Hi I am habeeb,
I am just finish Training in Bw,I am attending interview at the movement.I want know some real time naming conversion in realtime in SD,MM process.If you have functional spec's please send me documents my mail id.my mail id is: [email protected]
take care
bye
habeeb,Uk

Hello
The following are the steps involved in physical inventory
1.We prepare a list of the materials to be counted with material code/identification codes.
2.The inbound and outbound will be blockled at the time of counting.
3.Then we start the counting.
4.We enter the counted stock against the materials in the list.
5.Compare the counted physical stock with book stock or system stock.
6.If there is any differance then we will adjust the physical stock and book stock With permission from higher authorities.
Regards
Gregory Mathews

Similar Messages

  • Real time sample rate conversion?

    Recently upgraded all my hardware and software.. I want to record audio at 96khz - so I set my hardware to 96 and set logic to 96 in audio - no problem... except when I want to record new audio on an old project - I set the software to 96khz, and hey presto my old files recorded at 44.1 play too fast....
    The ref manual says logic will do real time sample rate conversion, but doesn't suggest how this is done.
    If I have my hardware set to 96khz but Logic to 44.1 so the old audio plays okay will my new audio still be recorded at 96?
    I could probably convert all the files individually using the sample rate converter in Factory, but I want to avoid this lengthy process as I'm talking the last 8 years of work!
    Anyone out there know anything about this?

    Just fancied recording at the best quality possible - but you are right, the difference won't be that noticable and may even sound out of place.... in any case I can record at 24 bit with no adverse effect...
    I noticed that trying to record audio at 96 when my hardware is 96 and logic at 44.1 has the effect of serious latency issues plus disk too slow errors. So forget that one.
    Someone told me recently that the difference in bit resolution is more audibly noticable as a change in audio quality than sample rate... just out of interest I wonder if anyone knows if this is true?
    BTW just recording guitars and stuff, some vocals, band kinda stuff...

  • How often in the real time projects extract data from flat files n process

    I am going thru teh BODS data integrator, and trying to understand the demand of ETL services extract data from a flat file, is that really impt in teh real time jobs.
    Thank you very much for the helpful info.

    Hi,
    As per the inputs given by you guys i started loading data from flat file.
    I try to load 28 files from i which i was able to load 24 files succesfully.For the other 4 i got this error messages
    1) Error 'Enter period in the format __.YYYY...' at conversion exit CONVERSION_EXIT_PERI6_INPUT (field CALMONTH record 1, value DUMYTRA)
    Message no. RSDS012
    2)  a) Error 'The argument '1,008.00' cannot be interpreted as anumber' on assignment field QUANT_B record 11714 value 1,008.00
    Message no. RSDS013
       b) Error 'The argument '1,110.00' cannot be interpreted as anumber' on assignment field QUANT_B record 15374 value 1,110.00
    Message no. RSDS013
    3) a) Error 'The argument '1,140.00' cannot be interpreted as anumber' on assignment field QUANT_B record 1647 value 1,140.00
    Message no. RSDS013
       b) Error 'The argument '2,028.00' cannot be interpreted as anumber' on assignment field QUANT_B record 4625 value 2,028.00
    Message no. RSDS013
    4) Error 'The argument '1,151.00' cannot be interpreted as anumber' on assignment field QUANT_B record 7808 value 1,151.00
    Message no. RSDS013
    I'am unable to trace out what is the error exactly.
    I checked this values in files they are perfect.
    can anybody please guide me on this issue.
    With Regards,
    Pradeep.B

  • Abap-hr real time questions

    hi friends
    kindly send me ABAP-HR REAL TIME QUESTION to my mail [email protected]
    Thanks&Regards
    babasish

    Hi
    Logical database
    A logical database is a special ABAP/4 program which combines the contents of certain database tables. Using logical databases facilitates the process of reading database tables.
    HR Logical Database is PNP
    Main Functions of the logical database PNP:
    Standard Selection screen
    Data Retrieval
    Authorization check 
    To use logical database PNP in your program, specify in your program attributes.
    Standard Selection Screen
    Date selection
    Date selection delimits the time period for which data is evaluated. GET PERNR retrieves all records of the relevant infotypes from the database.  When you enter a date selection period, the PROVIDE loop retrieves the infotype records whose validity period overlaps with at least one day of this period.
    Person selection
    Person selection is the 'true' selection of choosing a group of employees for whom the report is to run.
    Sorting Data
    · The standard sort sequence lists personnel numbers in ascending order.
    · SORT function allows you to sort the report data otherwise. All the sorting fields are from infotype 0001.
    Report Class
    · You can suppress input fields which are not used on the selection screen by assigning a report class to your program.
    · If SAP standard delivered report classes do not satisfy your requirements, you can create your own report class through the IMG.
    Data Retrieval from LDB
    1. Create data structures for infotypes.
        INFOTYPES: 0001, "ORG ASSIGNMENT
                            0002, "PERSONAL DATA
                            0008. "BASIC PAY
    2. Fill data structures with the infotype records.
        Start-of-selection.
             GET PERNR.
        End-0f-selection. 
        Read Master Data
    Infotype structures (after GET PERNR) are internal tables loaded with data.
    The infotype records (selected within the period) are processed sequentially by the PROVIDE - ENDPROVIDE loop.
              GET PERNR.
                 PROVIDE * FROM Pnnnn BETWEEN PN/BEGDA AND PN/ENDDA
                        If Pnnnn-XXXX = ' '. write:/ Pnnnn-XXXX. endif.
                 ENDPROVIDE.
    Period-Related Data
    All infotype records are time stamped.
    IT0006 (Address infotype)
    01/01/1990   12/31/9999  present
              Which record to be read depends on the date selection period specified on the
              selection screen. PN/BEGDA PN/ENDDA.
    Current Data
    IT0006 Address  -  01/01/1990 12/31/9999   present
    RP-PROVIDE-FROM-LAST retrieves the record which is valid in the data selection period.
    For example, pn/begda = '19990931'    pn/endda = '99991231'
    IT0006 subtype 1 is resident address
    RP-PROVIDE-FROM-LAST P0006 1 PN/BEGDA PN/ENDDA.
    Process Infotypes
    RMAC Modules - RMAC module as referred to Macro, is a special construct of ABAP/4 codes. Normally, the program code of these modules is stored in table 'TRMAC'. The table key combines the program code under a given name. It can also be defined in programs.The RMAC defined in the TRMAC can be used in all Reports. When an RMAC is changed, the report has to be regenerated manually to reflect the change.
    Reading Infotypes - by using RMAC (macro) RP-READ-INFOTYPE
              REPORT ZHR00001.
              INFOTYPE: 0002.
              PARAMETERS: PERNR LIKE P0002-PERNR.
              RP-READ-INFOTYPE PERNR 0002 P0002 .
              PROVIDE * FROM P0002
                  if ... then ...endif.
              ENDPROVIDE.
    Changing Infotypes - by using RMAC (macro) RP-READ-INFOTYPE. 
    · Three steps are involved in changing infotypes:
    1. Select the infotype records to be changed;
    2. Make the required changes and store the records in an alternative table;
    3. Save this table to the database;
    The RP-UPDATE macro updates the database. The parameters of this macro are the OLD internal table containing the unchanged records and the NEW internal table containing the changed records. You cannot create or delete data. Only modification is possible.
    INFOTYPES: Pnnnn NAME OLD,
    Pnnnn NAME NEW.
    GET PERNR.
        PROVIDE * FROM OLD
               WHERE .... = ... "Change old record
               *Save old record in alternate table
               NEW = OLD.
        ENDPROVIDE.
        RP-UPDATE OLD NEW. "Update changed record
    Infotype with repeat structures
    · How to identify repeat structures.
    a. On infotype entry screen, data is entered in table form.
        IT0005, IT0008, IT0041, etc.
    b. In the infotype structure, fields are grouped by the same name followed by sequence number.
        P0005-UARnn P0005-UANnn P0005-UBEnn
        P0005-UENnn P0005-UABnn
    Repeat Structures
    · Data is entered on the infotype screen in table format but stored on the database in a linear  
      structure.
    · Each row of the table is stored in the same record on the database.
    · When evaluating a repeat structure, you must define the starting point, the increment and the
      work area which contains the complete field group definition.
    Repeat Structures Evaluation (I)
    · To evaluate the repeat structures
       a. Define work area.
           The work area is a field string. Its structure is identical to that of the field group.
       b. Use a DO LOOP to divide the repeat structure into segments and make it available for  
           processing in the work area, one field group (block) at a time.
    Repeat Structures Evaluation(II)
    Define work area
    DATA: BEGIN OF VACATION,
                  UAR LIKE P0005-UAR01, "Leave type
                  UAN LIKE P0005-UAN01, "Leave entitlement
                  UBE LIKE P0005-UBE01, "Start date
                  UEN LIKE P0005-UEN01, "End date
                  UAB LIKE P0005-UAB01, "Leave accounted
               END OF VACATION.
    GET PERNR.
         RP-PROVIDE-FROM-LAST P0005 SPACE PN/BEGDA PN/ENDDA.
         DO 6 TIMES VARYING VACATION
                 FROM P0005-UAR01 "Starting point
                     NEXT P0005-UAR02. "Increment
                 If p0005-xyz then ... endif.
          ENDDO.
    Processing 'Time Data'.
    · Dependence of time data on validity period
    · Importing time data
    · Processing time data using internal tables
    Time Data and Validity Period
    · Time data always applies to a specific validity period.
    · The validity periods of different types of time data are not always the same as the date selection period specified in the selection screen.
    Date selection period |----
    |
    Leave |----
    |
    · PROVIDE in this case is therefore not used for time infotypes.
    Importing Time Data
    · GET PERNR reads all time infotypes from the lowest to highest system data, not only those within the date selection period.
    · To prevent memory overload, add MODE N to the infotype declaration. This prevents the logical database from importing all data into infotype tables at GET PERNR.
    · Use macro RP-READ-ALL-TIME-ITY to fill infotype table.
    INFOTYPES: 2001 MODE N.
    GET PERNR.
        RP-READ-ALL-TIME-ITY PN/BEGDA PN/ENDDA.
        LOOP AT P0021.
             If P0021-XYZ = ' '. A=B. Endif.
        ENDLOOP.
    Processing Time Data
    · Once data is imported into infotype tables, you can use an internal table to process the interested data.
    DATA: BEGIN OF ITAB OCCURS 0,
                  BUKRS LIKE P0001-BUKRS, "COMPANY
                  WERKS LIKE P0001-WERKS, "PERSONNEL AREA
                  AWART LIKE P2001-AWART, "ABS./ATTEND. TYPE
                  ASWTG LIKE P2001-ASWTG, "ABS./ATTEND. DAYS
               END OF ITAB.
    GET PERNR.
    RP-PROVIDE-FROM-LAST P0001 SAPCE PN/BEGDA PN/ENDDA.
    CLEAR ITAB.
    ITAB-BUKRS = P0001-BURKS. ITAB-WERKS = P0001-WERKS.
    RP-READ-ALL-TIME-ITY PN/BEGDA PN/ENDDA.
    LOOP AT P2001.
          ITAB-AWART = P2001-AWART. ITAB-ASWTG = P2001-ASWTG.
          COLLECT ITAB. (OR: APPEND ITAB.)
    ENDLOOP.
    Database Tables in HR
    ·  Personnel Administration (PA) - master and time data infotype tables (transparent tables).
       PAnnnn: e.g. PA0001 for infotype 0001
    ·  Personnel Development (PD) - Org Unit, Job, Position, etc. (transparent tables).
       HRPnnnn: e.g. HRP1000 for infotype 1000
    ·  Time/Travel expense/Payroll/Applicant Tracking data/HR work areas/Documents (cluster  
       PCLn: e.g. PCL2 for time/payroll results.
    Cluster Table
    · Cluster tables combine the data from several tables with identical (or almost identical) keys
      into one physical record on the database.
    . Data is written to a database in compressed form.
    · Retrieval of data is very fast if the primary key is known.
    · Cluster tables are defined in the data dictionary as transparent tables.
    · External programs can NOT interpret the data in a cluster table.
    · Special language elements EXPORT TO DATABASE, IMPORT TO DATABASE and DELETE
      FROM DATABASE are used to process data in the cluster tables.
    PCL1 - Database for HR work area;
    PCL2 - Accounting Results (time, travel expense and payroll);
    PCL3 - Applicant tracking data;
    PCL4 - Documents, Payroll year-end Tax data
    Database Tables PCLn
    · PCLn database tables are divided into subareas known as data clusters.
    · Data Clusters are identified by a two-character code. e.g RU for US payroll result, B2 for
      time evaluation result...
    · Each HR subarea has its own cluster.
    · Each subarea has its own key.
    Database Table PCL1
    · The database table PCL1 contains the following data areas:
      B1 time events/PDC
      G1 group incentive wages
      L1 individual incentive wages
      PC personal calendar
      TE travel expenses/payroll results
      TS travel expenses/master data
      TX infotype texts
      ZI PDC interface -> cost account
    Database Table PCL2
    · The database table PCL2 contains the following data areas:
      B2 time accounting results
      CD cluster directory of the CD manager
      PS generated schemas
      PT texts for generated schemas
      RX payroll accounting results/international
      Rn payroll accounting results/country-specific ( n = HR country indicator )
      ZL personal work schedule
    Database Table PCL3
    · The database table PCL3 contains the following data areas:
      AP action log / time schedule
      TY texts for applicant data infotypes
    Data Management of PCLn
    · The ABAP commands IMPORT and EXPORT are used for management of read/write to
      database tables PCLn.
    · A unique key has to be used when reading data from or writing data to the PCLn.
      Field Name KEY Length Text
      MANDT X 3 Client
      RELID X 2 Relation ID (RU,B2..)
      SRTFD X 40 Work Area Key
      SRTF2 X 4 Sort key for dup. key
    Cluster Definition
    · The data definition of a work area for PCLn is specified in separate programs which comply  
       with fixed naming conventions.
    · They are defined as INCLUDE programs (RPCnxxy0). The following naming convention applies:
       n = 1 or 2 (PCL1 or PCL2)
       xx = Relation ID (e.g. RX)
       y = 0 for international clusters or country indicator (T500L) for different country cluster
    Exporting Data (I)
    · The EXPORT command causes one or more 'xy' KEY data objects to be written to cluster xy.
    · The cluster definition is integrated with the INCLUDE statement.
    REPORT ZHREXPRT.
    TABLES: PCLn.
    INCLUDE: RPCnxxy0. "Cluster definition
    Fill cluster KEY
    xy-key-field = .
    Fill data object
    Export record
    EXPORT TABLE1 TO DATABASE PCLn(xy) ID xy-KEY.
       IF SY-SUBRC EQ 0.
           WRITE: / 'Update successful'.
       ENDIF.
    Exporting Data (II)
    . Export data using macro RP-EXP-Cn-xy.
    · When data records are exported using macro, they are not written to the database but to a  
      main memory buffer.
    · To save data, use the PREPARE_UPDATE routine with the USING parameter 'V'.
    REPORT ZHREXPRT.
    *Buffer definition
    INCLUDE RPPPXD00. INCLUDE RPPPXM00. "Buffer management
    DATA: BEGIN OF COMMON PART 'BUFFER'.
    INCLUDE RPPPXD10.
    DATA: END OF COMMON PART 'BUFFER'.
    RP-EXP-Cn-xy.
    IF SY-SUBRC EQ 0.
        PERFORM PREPARE_UPDATE USING 'V'..
    ENDIF.
    Importing Data (I)
    · The IMPORT command causes data objects with the specified key values to be read from
       PCLn.
    · If the import is successful, SY-SUBRC is 0; if not, it is 4.
    REPORT RPIMPORT.
    TABLES: PCLn.
    INCLUDE RPCnxxy0. "Cluster definition
    Fill cluster Key
    Import record
    IMPORT TABLE1 FROM DATABASE PCLn(xy) ID xy-KEY.
       IF SY-SUBRC EQ 0.
    Display data object
       ENDIF.
    Importing data (II)
    · Import data using macro RP-IMP-Cn-xy.
    · Check return code SY-SUBRC. If 0, it is successful. If 4, error.
    · Need include buffer management routines RPPPXM00
    REPORT RPIMPORT.
    *Buffer definition
    INCLUDE RPPPXD00.
    DATA: BEGIN OF COMMON PART 'BUFFER'.
    INCLUDE RPPPXD10.
    DATA: END OF COMMON PART 'BUFFER'.
    *import data to buffer
    RP-IMP-Cn-xy.
    *Buffer management routines
    INCLUDE RPPPXM00.
    Cluster Authorization
    · Simple EXPORT/IMPORT statement does not check for cluster authorization.
    · Use EXPORT/IMPORT via buffer, the buffer management routines check for cluster
      authorization.
    Payroll Results (I)
    · Payroll results are stored in cluster Rn of PCL2 as field string and internal tables.
      n - country identifier.
    · Standard reports read the results from cluster Rn. Report RPCLSTRn lists all payroll results;
      report RPCEDTn0 lists the results on a payroll form.
    Payroll Results (II)
    · The cluster definition of payroll results is stored in two INLCUDE reports:
      include: rpc2rx09. "Definition Cluster Ru (I)
      include: rpc2ruu0. "Definition Cluster Ru (II)
    The first INCLUDE defines the country-independent part; The second INCLUDE defines the country-specific part (US).
    · The cluster key is stored in the field string RX-KEY.
    Payroll Results (III)
    · All the field string and internal tables stored in PCL2 are defined in the ABAP/4 dictionary. This
      allows you to use the same structures in different definitions and nonetheless maintain data
      consistency.
    · The structures for cluster definition comply with the name convention PCnnn. Unfortunately, 
       'nnn' can be any set of alphanumeric characters.
    *Key definition
    DATA: BEGIN OF RX-KEY.
         INCLUDE STRUCTURE PC200.
    DATA: END OF RX-KEY.
    *Payroll directory
    DATA: BEGIN OF RGDIR OCCURS 100.
         INCLUDE STRUCTURE PC261.
    DATA: END OF RGDIR.
    Payroll Cluster Directory
    · To read payroll results, you need two keys: pernr and seqno
    . You can get SEQNO by importing the cluster directory (CD) first.
    REPORT ZHRIMPRT.
    TABLES: PERNR, PCL1, PCL2.
    INLCUDE: rpc2cd09. "definition cluster CD
    PARAMETERS: PERSON LIKE PERNR-PERNR.
    RP-INIT-BUFFER.
    *Import cluster Directory
       CD-KEY-PERNR = PERNR-PERNR.
    RP-IMP-C2-CU.
       CHECK SY-SUBRC = 0.
    LOOP AT RGDIR.
       RX-KEY-PERNR = PERSON.
       UNPACK RGDIR-SEQNR TO RX-KEY-SEQNO.
       *Import data from PCL2
       RP-IMP-C2-RU.
       INLCUDE: RPPPXM00. "PCL1/PCL2 BUFFER HANDLING
    Function Module (I)
      CD_EVALUATION_PERIODS
    · After importing the payroll directory, which record to read is up to the programmer.
    · Each payroll result has a status.
      'P' - previous result
      'A' - current (actual) result
      'O' - old result
    · Function module CD_EVALUATION_PERIODS will restore the payroll result status for a period
       when that payroll is initially run. It also will select all the relevant periods to be evaluated.
    Function Module (II)
    CD_EVALUATION_PERIODS
    call function 'CD_EVALUATION_PERIODS'
         exporting
              bonus_date = ref_periods-bondt
              inper_modif = pn-permo
              inper = ref_periods-inper
              pay_type = ref_periods-payty
              pay_ident = ref_periods-payid
         tables
              rgdir = rgdir
              evpdir = evp
              iabkrs = pnpabkrs
         exceptions
              no_record_found = 1.
    Authorization Check
       Authorization for Persons
    ·  In the authorization check for persons, the system determines whether the user has the 
       authorizations required for the organizational features of the employees selected with
       GET PERNR.
    ·  Employees for which the user has no authorization are skipped and appear in a list at the end
       of the report.
    ·  Authorization object: 'HR: Master data'
    Authorization for Data
    · In the authorization check for data, the system determines whether the user is authorized to
      read the infotypes specified in the report.
    · If the authorization for a particular infotype is missing, the evaluation is terminated and an error
      message is displayed.
    Deactivating the Authorization Check
    · In certain reports, it may be useful to deactivate the authorization check in order to improve
      performance. (e.g. when running payroll)
    · You can store this information in the object 'HR: Reporting'.
    these are the main areas they ask q?

  • Video Output for PXI Embedded Real-Time Controller ?

    I plan to purchase a PXI Embedded Real-Time Controller for security monitoring.
    Is it possible to use the video output of a PXI Controller running RTOS to display messages from a LabVIEW RT application ? According to the documentation, it seems that the use of the video output is only possible for PXI Controllers running Windows XP.
    Does it exist a specific application to do this with LabVIEW RT ?
    Thanx.
    H.L.
    Solved!
    Go to Solution.

    Hi Yann,
    The situation seems a bit confusing as in the link you've given above using the local monitor from a pxi labview rt system as the HMI without a pc attached, two messages (here and here) say that it is possible to display an image on the video output of a PXI RT controller using a function of the NI-IMAQ librarey (namely, the IMAQ RT Video Out VI) !
    I have the NI-IMAQ library but no PXI controller to test if this solution should work or not
    Does anyone have the answer ?
    Thanx,
    H.L.

  • How to calculate rms voltage, rms current and frequency from real time data?

    Hello,
    I need to calculate the real time rms voltage, rms current and frequency of the received voltage and current graphs using visa read. Actually am trying to monitor the voltage and current parameters of AC load, in this case, am using a set of 4 light bulbs at the moment. Expected frequency is 50 Hz, and rms voltage close to 240v AC, current about 1A. 
    Attached is the vi picture of what I have been able to achieve so far, courtesy of labview forums. But I need to finally finish this vi by calculating the Vrms, Irms and frequecy. The front panel shows the voltage (Chart V) and current (Chart I) after scaling. Chart VI represents ADC value 0 ~ 1023 of received data. Aray V & I show the binary bits received up to 10 bits of useful data from ADC. Each sine wave from ADC is sampled at 25 samples for the ADC conversion, and sampling frequency of 1.28kH.
    I have tried using the RMS vi, but seems I don't know how to configure it correctly to get desired results. Also when I make changes like switching of one of the bulbs so that their is current change, I need to refresh the SERIAL port first before changes show up on the VI. Any ideas on how to improve on this are highly appreciated.
    Otherwise, I would appreciate if somebody helps me MODIFY my VI to read the rms values and frequency from received data. This is the last piece of my project, I do appreciate all the help rendered. Am currently using labview labview 2014, student version!
    Gavin.
    Attachments:
    Test_revised1.vi ‏39 KB
    Test_frontpanel.png ‏196 KB
    Test_blockdiagram.png ‏187 KB

    hello,
    note that peak voltage (Vpeak) is (240 * 1.414) = 399 V and peak current about 1.414 if intended rms current is 1A.
    Gavin.

  • Real Time Integration Document - Rounding Error between LC1 and LC2

    Hi!
    We recently underwent currency conversion. After the conversion, we noticed that Real Time Integration document that is generated due to Assessment Run (KSU5) produces a document which are off by cents between LC1 and LC2 ,even though both have the same currency
    LC1 = EUR (Currency Type 10 - Company Code Currency)
    LC2 = EUR (Currency Type 60 - Global Company Currency)
    Document being generated
    Document Currency = USD (Controlling Area Currency)
    Account     LC1 (EUR)      LC2 (EUR)
    XXX            2297.44          2297.43
    Any help would be appreciated .

    Hello Manish,
    I am encountering the same issue. Have you found solution on this.
    BR,
    Zulfikar

  • Plz Give me real time file to file scenarios in sap xi

    Hello Experts,
       Can any one give me the real time file2file scenarios in sap xi. And differnet file structures
    and different 'File Content Conversion'  objects. Explain me different approch for file2file with different scenarios.
    Thanks in advance.
    Arjun.

    HI Arjun,
    Go throgh the following links for File content conversion in the sender side as well as receiver side
    File content conversion sites:
    Introduction to simple(File-XI-File)scenario and complete walk through for starters(Part1)
    Introduction to simple (File-XI-File)scenario and complete walk through for starters(Part2)
    NAB the TAB (File Adapter)
    Introduction to simple (File-XI-File)scenario and complete walk through for starters(Part2)
    How to send a flat file with various field lengths and variable substructures to XI 3.0
    Content Conversion (Pattern/Random content in input file)
    Content Conversion ( The Key Field Problem )
    File Receiver with Content Conversion
    http://help.sap.com/saphelp_nw04/helpdata/en/d2/bab440c97f3716e10000000a155106/content.htm
    Please see the below links for file content conversion..
    The specified item was not found. - FCC
    File Content Conversion for Unequal Number of Columns
    File Content Conversion for Unequal Number of Columns - FCC
    Content Conversion (Pattern/Random content in input file)
    Content Conversion (Pattern/Random content in input file) - FCC
    XI Configuration for MDM Integration - Sample Scenario - FCC - MDM
    Regards
    Sridhar Goli

  • Share small real-time example using Narrowing Cast & Widening Cast

    Hi Guys,
    I am trying to understand the concept of Narrowing Cast and Widening Cast and getting confused why we are using these Casts. When we can always access the method of Subclass using its ref object in Narrowing ) and use the Super class with its ref object (in Widening).
    I understand the concept but could not understand under what situation cast is preferred.There are articles all over the web to explain the concept, but non told us why and when we are forced to us these CASTS.
    I will appreciate if somebody can share a small real-time example where use of Narrowing and Widening Cast is advisable.
    Thanks,

    It looks like this topic has lost steam. 
    I'm surprised this isn't a more intense conversation. There are many applications that need to share data.
    In my application I have many hundreds (in some few cases in the thousands) of different types of data streams including DIO, Analog, and many different types of Serial buses.
    My application is already controlling the hardware and data streams, but others want to peak into my I/O including what is happening on the serial buses.
    To create a 'get' function for each I/O in a web page would seem farily intense (although I could probably create a VI script to help build it).
    I was looking for solutions along the lines of maybe coping the data into a VISA service (if that is possible), or publishing to a datasocket server, or creating a network published variable.
    I need to be considerate of the possibility of hundreds of serial buses banging a service and the possible bandwidth impacts. So I'm searching for a solution that is both easy to implement and low on resources.
    For now I'm leaning towards datasockets. Does anybody have an opinions on this?

  • About DTP with real-time access.

    Hello Gurus,
           setting for error handling only has an impact while repairing a DTP request (repair) and during the conversion of the DTP to standard DTP (for example, to correct an error during extraction).
          will you please give a simpel scenario to explain above words?
    thank you very much.
    Fran

    Hi,
    Use :
    With SAP NetWeaver 7.0, SPS 14, the following changes and enhancements are available for real-time data acquisition:
    ●      Changes to the menu and the context menu of the monitor for real-time data acquisition
    The menu and the context menus for the individual objects in the monitor for real-time data acquisition have been standardized and enhanced. In particular, it is now possible to assign daemons and data transfer processes on various levels using context menu entries. For a complete overview of the functions in the menu and in the context menus, see Monitor for Real-Time Data Acquisition.
    ●      Assignment of daemons for InfoPackages and data transfer processes (DTP)
    To assign InfoPackages and data transfer processes to a daemon, you can call the monitor for real-time data acquisition in the following ways:
    ○       In the Data Warehousing Workbench using the respective context menu entry Assign RDA Daemon.
    ○       In InfoPackage maintenance using the Schedule tab page, and in data transfer process maintenance using the Execute tab page.
    The button names for jumping to the monitor for real-time data acquisition have changed here. To jump to the monitor, choose Assign Daemon.
    ●      Repair process chains for repairing a broken data transfer or a missing delta in the DataStore object
    In certain situations, a gap in the delta of the DataStore object can occur if there is a closed request in the PSA but there is no corresponding request in the DataStore object. For example, this is the case if a DTP request has terminated due to an error in the transformation. Here, you can create a repair process chain to repair a missing or broken update from the PSA. The repair process chain contains a standard DTP as well as any further processes required for subsequent processing (such as activating the data in the DataStore object or subsequent process chains). When the repair process chain is executed, the complete delta is loaded from the source into the DataStore object and processes for activation and further processing are executed if required.
    ●      Process types for starting and stopping real-time data acquisition
    You can use process chains to control real-time data acquisition using process types Start Real-Time Data Acquisition (RDA) Load Process and Stop Real-Time Data Acquisition (RDA) Load Process.
    More Info :
    http://help.sap.com/saphelp_nw70/helpdata/en/47/2731751c2a2dede10000000a1553f7/frameset.htm
    Regards
    Ram.

  • Real time : File Adapter : query

    Hi Folks,
    At development time( file adapter) we test our scenario by creating a sample XML file using message mapping tool .
    The type of message we get from this contains Message Type tag with namespace Attribute.
    <ns0:MT_XYZ xmlns:ns0="urn:xyz">
    My question is in do in real time we need to tell them that create message type with the anmespace attribute we are using ? and also how to suppress namespace and will it bring any complications in the process ?
    Regards,

    Santosh,
    1.     When using the File adapter and Content Conversion to read the file into XI, you can allocate a specific message namespace
    2.     If the namespace does becomes an issue for your scenario then use an additional mapping step to remove the namespace.  For example we have the
    Mesage Type MT_NAMES which has a a type of NAMES
    Subelement NAMEREC (0 .. unbounded)
    Subelements NAME1 and NAME2
    Sample xml
    <?xml version="1.0" encoding="UTF-8" ?>
    <ns1:MT_NAMES xmlns:ns1="http://namespace">
    <NAMEREC>
    <NAME1>AAA</NAME1>
    <NAME2>BBB</NAME2>
    </NAMEREC>
    <NAMEREC>
    <NAME1>CCC</NAME1>
    <NAME2>DDD</NAME2>
    </NAMEREC>
    </ ns1:MT_NAMES>
    Apply the XSL map
    <?xml version="1.0" encoding="ISO-8859-1"?>
    <xsl:stylesheet version="1.0"
    xmlns:xsl="http://www.w3.org/1999/XSL/Transform"     
    xmlns:nsNN="http://namespace" >
         <xsl:template match="/">
              <xsl:for-each select="nsNN:MT_NAMES">
                   <NAMES>
                        <xsl:copy-of select="NAMEREC"/>
                   </NAMES>
              </xsl:for-each>          
         </xsl:template>
    </xsl:stylesheet>
    The namespace has been removed from the xml
    <?xml version="1.0" encoding="utf-8" ?>
    <NAMES>
    <NAMEREC>
    <NAME1>AAA</NAME1>
    <NAME2>BBB</NAME2>
    </NAMEREC>
    <NAMEREC>
    <NAME1>CCC</NAME1>
    <NAME2>DDD</NAME2>
    </NAMEREC>
    </NAMERS>
    Regards,
    Mike

  • How to create Real Time Infocube in BW

    Hi BW Experts,
    We got one requirement. We need to create one Real Time Infocube.
    Our Users will enter the data in Query Analyzer later.
    Please can anyone tell me the step by step.
    Is there any settings we need to give..
    Thanks & Regards
    Anjali

    Hi,
    Creating a Real-Time InfoCube
    When creating a new InfoCube in the Data Warehousing Workbench, select the Real-Time indicator.
    Converting a Standard InfoCube into a Real-Time InfoCube
    Conversion with Loss of Transaction Data
    If the standard InfoCube already contains transaction data that you no longer need (for example, test data from the implementation phase of the system), proceed as follows:
           1.      In the InfoCube maintenance in the Data Warehousing Workbench, from the main menu, choose InfoCube ® Delete Data Content. The transaction data is deleted and the InfoCube is set to inactive.
           2.      Continue with the same procedure as with creating a real-time InfoCube.
    Conversion with Retention of Transaction Data
    If the standard InfoCube already contains transaction data from the production operation that you still need, proceed as follows:
    Execute ABAP report SAP_CONVERT_NORMAL_TRANS under the name of the corresponding InfoCube. Schedule this report as a background job for InfoCubes with more than 10,000 data records because the runtime could potentially be long.
    Try
    http://help.sap.com/saphelp_nw04s/helpdata/en/39/100c38e15711d4b2d90050da4c74dc/frameset.htm
    Hope this helps.
    Thanks,
    JituK

  • LabVIEW 8.5 & Real Time

    Hello..
    I'm a new comer in LabVIEW... and till now i was working with LabVIEW 7.1 for my projects.. but now onwards i need to work in LabVIEW 8.5 for my next project... So can anyone please tell me where can i get evaluation version  of LabVIEW 8.5 and LabVIEW Real Time... so that i can practice with it...
    Thanks in advance..
    Solved!
    Go to Solution.

    Before you install the new version of LabVIEW, copy the folder structure from C:\Program Files\National Instruments\LabVIEW 7.0\ to a separate folder, drive, or CD.  Copy everything in the folder.
    Install the new version of LabVIEW.  The new installation will remove or modify drivers used by LabVIEW 7.
    When the installation of LabVIEW 8.x completes, restore the folder structure for LabVIEW 7 from your backup copy.  Don't make any changes to the new folder for LabVIEW 8.x.  Both versions will be available.
    Cautions:
    1) Do not run both versions simultaneously.
    2) Be very careful about converting your old code.  If you convert anything, rename it and all sub-VIs associated with it.  Copy it to a new folder before converting it.
    Using this method, I have a computer with 4  or 5 versions of LabVIEW peacefully coexisting.  Just remember to think about conversions before doing them.  It's hard to recover from inadvertent conversion.

  • Liquid Analytics enables the sales force in real-time

    Problem
    The success of wholesale and manufacturing distribution operations largely hinges on the efficacy of sales teams that must deal with a host of pricing and product challenges. For the sales group, the biggest game changer is their ability to access and leverage insight into the customer’s behavior and historical buying patterns.
    Due to the technological limitations of the past, the sales team was unable to access information about customers’ buying patterns, leverage new products, establish dynamic pricing, and set realistic goals and as a result would miss opportunities with customers or lose their faith altogether.
    Even with these pressing needs noted and observed, not just any technology can be introduced to a sales team and expect to be adopted. This is because the unique characteristics of the sales force must be carefully considered in order to seamlessly integrate new technologies aimed at sales enablement.
    Liquid analytics understands the field and the interplay between marketers, sales teams, and customers and uses this knowledge to inform its technology. The firm understands that the sales force operates primarily in the field and therefore does not have the capability to manually input data into sales tracking forms, or that sales teams with different products visit the same customers but do not necessarily work in tandem to leverage each other’s footprints. Without a foolproof means of measurement and assessment in the field, the disconnect between the sales and marketing teams is often heightened making set goals unrealistic for sale representatives.
    Solution
    Liquid analytics uses its understanding of the sales archetype and big data and predictive analytics to bring viability to its target industries when it counts most. The proprietary technology takes predictive algorithms and mobile tracking, visualizing the various components in a comprehensive, easy-to-use dashboard. For mobile sales representatives, valuable information can be easily searched on an account, and sales activity is easily logged to demonstrate ROI.
    For wine and liquor sales representatives, the solution has been proven to do a number of things:
    The technology uses algorithms based on history of sales which are able to forecast what will be needed and allows the sales team to make fact- based orders, save time on each customer account by avoiding duplication, and increase customer loyalty. In addition, access to inventory in real time allows the sales rep to deliver in a timely manner and avoid missed shipments by 95 percent.
    The Liquid analytics dashboard allows representatives selling liquor, wine or beer to view one universal dashboard to note what sales tasks have yet to be done or have been completed and develop a workflow that collectively benefits all representatives on the account.
    The leaderboard creates a gamification experience for the sales team meant to increase sales by incentivizing through potential conversion rates and goal reward system.
    On an individual basis, representatives are able to log their hours and use the record to negotiate better terms. Logged hours also provide context as to whether a representative is spending too much time while profitability remains low, leading to new sales strategies.
    The feedback loop between the marketing and sales team is closed as the solution generates real-time notifications about where media is used by sales team as well as categorizes product attributes that can be used for better product placement and targeted campaigns.
    Watch the Video
    Join the conversation and follow us on twitter @SAPstartups #HANAusecases.

    Thanks for this pointer.
    I followed this link and enabled 0fi_gl_4 for real time reporting in R/3
    What do I do now in BW side I already have a infsource that I am updating a DSO/ODS in BW. What do I need to do in BW side to ensure that the data in R/3 will flow through to BW.
    Thanks
    Karen

  • How to see the bitrate of a video in real time

    Some time ago I posted a question on this forum, asking if it's possible to do that (http://discussions.apple.com/thread.jspa?threadID=325516). As the topic is archived, I can not write the answer I found today: VLC, at least version 0.8.5, can do exactly what I need: on Window > Information > Statistics, it's a reading named Stream bitrate that reads like once per second the bitrate.
    It works not only for DVDs, but for any video, of course.
    Brilliant! And it's free!!

    This is great!
    You can both watch your newly compressed m2v and see the bitrate in real time.
    VLC media player 8.5 for Mac OS X can be downloaded here:
    http://www.videolan.org/vlc/download-macosx.html

Maybe you are looking for