Sample schema - real life data to make it interesting for stundents

Hi Guys,
I wonder if there is an exciting example training schema around and
I dont mean HR or customer order data :) I rather
want to see something that makes students want to do the join and
figure out the details. Data like salary statistics for
professions and various countries by year or other
interesting real-life demographic data.
Thanks for any ideas
metalray

Hi Centinul,
Thanks for the link.
@Pavan, I dont have any requirements. I just want a sample schema that does
not make stundents fall asleep while querying some random numbers (who cares about sales figures of john doe?).
I need to keep people engaged and make them "want" to query because they "want" to know the answer.

Similar Messages

  • Merging XML Data to make many documents

    Hello,
    Here's where I am at:
    I've used LifeCycle Designer to create a PDF with all the forms named. I have XML data for all my companys. The XML data merges into the document, and fills in the form perfectly.
    However, I'm looking to take this one step further, and automate the process. Meaning, Putting all the company data in one file, and importing the data to make PDF forms for each company.
    Similar to a mail merge. --Lots of different Companys, Fill in thier each individual page.
    What are the steps to do this?
    Thanks!
    --Jeff

    Yes.
    You can do one of the following:
    a) Build a custom application using the LiveCycle Forms api.
    b) Use the RenderForm QPAC within a workflow to convert the XML to a PDF.
    The RenderForm QPAC is a little obscure to use - read the documentation VERY carefully, and be very careful about the values you enter, otherwise it won't work. It does work perfectly if you get all the values right. In my experience, you do need to enter a username and password, even though these are not noted as mandatory fields.
    Howard
    http://www.avoka.com

  • Sampling Schema in QM

    How the sampling schema will work ? .
    We want to set the up the sampling schema. When we are recording the results for the usage decision, we want the sampling schema to influence the results. for example if the sampling is 10 % , means we want the user record 10 % sampling . How to do this function.
    Any business process steps with configuration  gudie will be highly appreciated .
    Thanks
    JJ

    What is the difference between the Qualitative chars and Qualitative chars in MIC
    Hi JJ,
    I think by this you mean to ask that what is the diff. between Qualitative Char & Quantitative Chars. Right?
    Qualitative Chars:-
    As it's name suggest itself the test that are performed to measure the qualitative analysis of the material, e.g. Color, Odour, Solubility, Description etc....For these all test, the qualitative Characteristics are maintained in SAP.
    The Test for which the results are likely,supposed to, expected to receive as Characters, Description, as  "TEXT", they are normally maintained as qualitative characters in SAP.
    Quantitative Chars:-
    As it's name suggest itself the test that are performed to measure the quantitative analysis of the material, e.g. Weight, Hight, Width etc....For these all test, the qualitative Characteristics are maintained in SAP.
    The Test for which the results are likely,supposed to, expected to receive as Numbers they are normally maintained as qualitative characters in SAP.
    Dear All SAP-PLM Champs....
    Herewith I have tried to elaborate it in very ground level words, if I am wrong somewhere pls correct me.
    Regards,
    Shyamal

  • How are Game Center "Friend Recommendations" people I know in real life? (Meetup/Facebook data?)

    This seems like a major privacy concern, but mainly I'm just very curious how this can possibly happen:
    The Game Center app suggests primarily people I have met in real life... but have had NO online interaction with!
    I life in a high-population US city, yet at least 2/3 of my Friend Recommendations (maybe more if they had photos or knew these peoples' last names) are people I have actually MET in person and know vaguely... but who I do now know well and who are NOT friends/contacts in any online service.
    I have never emailed them (with two exceptions); we don't even know each others' email addresses, and even if we did, my Apple ID email is not one I use for actual emailing purposes. Nor do we know each others' phone numbers, nor gaming nicknames; often not even last names. I didn't even know that they owned iOS devices or played games at all!
    We have certainly never played games with one another before; in fact, I have never used Game Center's friend feature and I don't have any GC friends. I only play random matches, and almost always on non-GC game services, for that matter.
    None of them has ever made a friend request to me. I've only ever gotten one GC request, long ago, and I denied it; it was not someone I knew.
    Some of them list "zero games in common" so it's not even that (and anyway, millions of people have games in common with me).
    Only one of them is in my Contacts (iOS/Mac address book), and knowing how serious they are about privacy, I doubt the email address I have is their AppleID. I know your address book can (for now) be accessed without permission, but these people just aren't listed.
    But I DO run into these recommendations in person every so often!
    These cannot be randomly selected by geography because a) they're not super close to me in miles, and b) there are millions of iOS users near me. No way that most of the suggestions would be people I've met, if this were chance alone. SOME source of data is being used by GC, and I'm baffled as to what it could be.
    The closest three theories I can come up with are all pretty implausible:
    1. Indirect Facebook connection. These people are NOT my Facebook friends (with a couple exceptions; I never actually use Facebook but I do have an account, and if someone requests, I generally accept). I double-checked that. But they MAY be "friends of friends" possibly. But how would Apple obtain my FB friend list? And if they did have it, wouldn't they just suggest my actual FB friends, not friends-of-friends? My Facebook login email address is an alumni address that I have never used for any other purpuse; it is not my Apple ID. And I have diligently set nearly every option to "hidden" on Facebook: even friends see a mostly blank profile!
    2. Meetup.com. This is a 100% correlation! I'm in several local groups that use Meetup.com to schedule meetings, and ALL of these Game Center recommendations are memebers of Meetup groups that I too am a member of. That's how I've met most of them. But Meetup has no "friends" feature, and most of these people have had no online connection with me (Meetup messages etc.) through Meetup. We've just attended the same events--which do have online RSVP lists--or else we are merely listed as members of the same group(s). (But we're talking large groups, so I haven't actually met most of the members, and I have met most of these GC recommendations.) But how would Apple obtain my Meetup info? Again my Meetup account's email address is not my Apple ID, and I have never used that email address with any Apple service (although it is on my Contacts card).
    3. Maybe Apple harvested my address book/contacts list (which does NOT contain these people), and also harvested the contacts lists of these people, AND also harvested the contacts lists of people we know in common who happen to have both of us listed. Then they put it all together. (That's a stretch: the people I know the best on Meetup--whose info I do have in Contacts--are Android users!) But again, why not recommend my actual contacts, instead of contacts-of-contacts? My actual contacts--which contain numerous iOS users and games--are notably absent from the recommendations (with one exception).
    Obivously Apple cannot, and does not, literally know who I have met. (No foil hat for me!) But they DO seem to have some data source that correlates closely with people I have met (and Meetup activity especially). I'm dying to know how that's possible.
    This is a privacy concern, for at least three reasons I can think of:
    1. GC lists these people's real names AND GC nicknames together--BEFORE I have accepted them as friends! And vice versa, I'm sure my nickname is being shown with my real name to total strangers. That means strangers, employers, stalkers, whoever, can now tie a person to their nickname (often the same one they use all over the Interner, not just GC). But we're not talking random suggestions: this is being show to people I DO know. Maybe I don't want them Googling my online nicknames! Maybe I complain about my work with that nickname Maybe I want my online friends and real-life acquaintainces kept separate. I definitely want EVERYTHING kept separate from these people I barely know but run into from time to time! One of these people is, in fact, on the sex offenders registry, and has stalked a friend of mine both online and offline. Maybe he'll learn my GC nickname--the same way it has shown me his--and can now Google that!
    2. If I have a lot of "games in common" listed, then they can peg me as a "gamer," which is sometimes a negative. Maybe I want my boss (or whoever) to think I'm more serious, even if he games too, and now he knows I have 30 games and wonders why I'm not staying late to work Maybe I just don't want to be roped into being "friends" online with my boss, or having to let him win games! Again, although this is a small issue at first glance, it's less small when things are being shared with people you really DO know in real life. What if my hypothetical stalker sees I have games in common? Sure, I can deny their friend request and cancel their 500 game attempts, but I'd rather be 100% hidden from them in the first place. I have no prior online contact, I want it to stay that way, and I've never given Apple permission, that I know of, to share my real name with people I never friended. (And even if I did, it should be with random strangers--not actual acquaintances.)
    3. The big mystery--how is Apple getting data that correlates with who I've met in the first place? I'm guessing that some people see the same pattern I do, and some don't; which might lead to an answer, but so far I'm stumped. (If this does tie to Meetup some bizarre way, that would be unnerving because Meetup ties me to real places and specific times.)
    Theories are welcome! And a basic question: what data sources DOES Game Center have available to use in making friend suggestions? I would have guessed maybe Location Services (nearby people) and people who have played a game with me (or made a friend request) in the past. But those sources are not what I'm seeing.
    (Even weirder is that nobody else seems to have posted about this happening; it may not happen to everyone, but surely some people have had this same question? I feel like I'm in a Kafka novel...)

    That's a good thought that could explain some cases, but not mine: these people definitely don't know my email address, and I'm not in their Address Book. (Some of these are people I might meet on a hike, say, and nod to in passing on other hikes once a or twice a year, and that's the extent of the "relationship"--no contact info ever exchanged. We don't even know what cities each other are from.)
    we don't even know each others' email addresses, and even if we did, my Apple ID email is not one I use for actual emailing purposes. Nor do we know each others' phone numbers, nor gaming nicknames; often not even last names.
    I should correct my first post: I said I "do now know well"... that should be "do NOT know well".

  • Sample schema and data downloads?

    I have installed oracle and I wish to test it with a substantial schema installed, with enough data to simulate a production level database.
    I was wondering if online there were any large sample schemas, with large amounts of data (in sql loader .dat files?) that could be downloaded. My search so far has not been successful.
    If possible I would like to get a schema with very large number of tables and many thousands of lines of data.

    First thing: you'd be better generating a dataset to match your particular requirements.
    However, If you're thinking about a data warehouse, you could do worse than borrow the Mondrian 'Foodmart' test dataset. That contains 328k rows of data (so pretty small, but not trivial). See http://mondrian.sourceforge.net/.
    Alternatively, lots of people start with datasets based on Northwind - the well known Access/SQL Server dataset. There's an article here http://www.oracle.com/technology/pub/articles/gagne_access.html on how to extract the necessary DDL using the mdbtools utilities.
    HTH
    Regards Nigel

  • Data Types from Oracle Sample Schema

    Hi,
    I was trying to push some data from an Oracle Sample schema (Schema: OE Table: Customers) to a target schema. The interface was erroring out because of undefined data types in the model.
    I chkd the model, only to find that some of the data types were undefined.
    I DESCribed the table in the using SQL Client, The data types returned were
    CUST_ADDRESS_TYP
    PHONE_LIST_TYP
    MDSYS.SDO_GEOMETRY
    I DESCribed the datatypes, They seem to be flexfields of some kind,
    Please advise on how to define these data types into ODI:
    Heres the DESC output
    SQL> desc customers;
    Name Null? Type
    CUSTOMER_ID NOT NULL NUMBER(6)
    CUST_FIRST_NAME NOT NULL VARCHAR2(20)
    CUST_LAST_NAME NOT NULL VARCHAR2(20)
    CUST_ADDRESS CUST_ADDRESS_TYP
    PHONE_NUMBERS PHONE_LIST_TYP
    NLS_LANGUAGE VARCHAR2(3)
    NLS_TERRITORY VARCHAR2(30)
    CREDIT_LIMIT NUMBER(9,2)
    CUST_EMAIL VARCHAR2(30)
    ACCOUNT_MGR_ID NUMBER(6)
    CUST_GEO_LOCATION MDSYS.SDO_GEOMETRY
    DATE_OF_BIRTH DATE
    MARITAL_STATUS VARCHAR2(20)
    GENDER VARCHAR2(1)
    INCOME_LEVEL VARCHAR2(20)
    SQL> desc CUST_ADDRESS_TYP;
    Name Null? Type
    STREET_ADDRESS VARCHAR2(40)
    POSTAL_CODE VARCHAR2(10)
    CITY VARCHAR2(30)
    STATE_PROVINCE VARCHAR2(10)
    COUNTRY_ID CHAR(2)
    SQL> desc phone_list_typ;
    phone_list_typ VARRAY(5) OF VARCHAR2(25)

    To create MDSYS.SDO_GEOMETRY datatype, goes to Topology module, Oracle Techonology, Datatypes, and add a new datatype GEOMETRY. (try to fill all information fields with GEOMETRY and see what happens maybe you have to ajust this after but it is a start point).
    add information in from and to tabs according your source and target desired tech.
    Reverse againg your model, or change the datatype in the model datastore. Re-run your interface

  • Real life SD Project - Gurus requested to give brief feedback

    Hi,
    Please read the real life SD - FICO scenario given below and if you think you have the solution, kindly state your suggestion. Would be much appreciated.
    ===================================================
    Engineering Services Orders  (Inter-company) 
    Process Steps  (As-is)
    1.Purchase Order is received in Co. A – via EDI,  from Co. B
    2.Sales Order (services) is automatically generated in Co. A. It has item 10 as hours and item 20 as travel expenses in $ amount
    Sales accounting creates ‘settlement rule’ and ‘internal order’ number and advises the engineers who will provide the services
    3.Later on, engineers report actual hours for services rendered, by using CATS system This merely reports data in SAP system but does not make an FI posting
    They use the order number advised to them in step 2 to create an internal order
    4 Actual travel expenses are added by Payroll  -  as a direct posting to FI
    5. Periodic CATS batch job is carried out by relevant IT person, to settle costs to an ‘internal order’ (not to sales order). These costs are labor hours and travel expenses
    6. FICO personnel then do transaction K08G to settle internal order --> sales order
    7. (Manual Step) Sales Accounting then runs report KVBI to view and select data related to service orders The data is downloaded to an Excel worksheet and handed over to SD-Billing.
    8. (Manual step) The cost data in dollar figures is converted to hours by SD-Billing, dividing cost data with rate / hour given in the sales order condition record
    9. SD-Billing creates invoice. Since it is order-related invoice, it has open quantities coming in from the order. These are now manually over-written / replaced with the actual data from step 7 & 8. An invoice is thus, generated for each order individually.
    Process (To-be)
    To create an interim document (billing request?) to replace manual steps 7 and 8. This interim document needs to capture actual cost data form the system –  and also show the remaining / open quantities from the sales order
    QUESTION: what is the best route to go?
    •  Configure a Billing Request document (by modifying a standard Credit Memo   Request / Debit Memo Request)  or
    • Going via the route of resource-related billing?

    Refer to the document: J48 - Project with Intercompany Billing
    Link: http://help.sap.com/bp_blv1500/BL_India/index.htm
    This is a very good document for guidance.
    Regards,
    Rajesh Banka

  • Can't find sample schemas

    I have installed 10g xe. I am self training using the oracle press book "Oracle Database 10g A Beginner's Guide." However, the book indicates that the customer table, mentioned in chapter 2 is provided as part of the sample schema in 10g. But the only schema I have is HR. Is there a place where I can download and import the schemas mentioned in the book? Or must I install enterprise edition in order to work with them? Thanks to any that can assist.

    The Order Entry (OE) schema builds on the purely relational Human Relations (HR) schema with some object-relational and object-oriented features. The OE schema contains seven tables: Customers, Product_Descriptions, Product_Information, Order_Items, Orders, Inventories, and Warehouses. The OE schema has links into the HR schema and PM schema. This schema also has synonyms defined on HR objects to make access transparent to users.
    rem
    rem Header: oe_cre.sql 09-jan-01
    rem
    rem Copyright (c) 2001, 2002, Oracle Corporation.  All rights reserved. 
    rem
    rem Owner  : ahunold
    rem
    rem NAME
    rem   oe_cre.sql - create OE Common Schema
    rem
    rem DESCRIPTON
    rem   Creates database objects. The script assumes that the HR schema
    rem   is present.
    rem
    rem NOTES
    rem   The OIDs assigned for the object types are used to
    rem   simplify the setup of Replication demos and are not needed
    rem   in most unreplicated environments.
    rem
    rem MODIFIED   (MM/DD/YY)
    rem   hyeh      08/29/02 - hyeh_mv_comschema_to_rdbms
    rem   ahunold   09/17/01 - FK in PRODUCT_DESCRIPTIONS
    rem   ahunold   04/25/01 - OID
    rem   ahunold   03/02/01 - eliminating DROP SEQUENCE
    rem   ahunold   01/30/01 - OE script headers
    rem   ahunold   01/24/01 - Eliminate extra lines from last merge
    rem   ahunold   01/05/01 - promo_id
    rem   ahunold   01/05/01 - NN constraints in product_descriptions
    rem   ahunold   01/09/01 - checkin ADE
    PROMPT
    PROMPT specify Sample Schema version as parameter 1:
    DEFINE vrs     = &1
    PROMPT
    -- ======================================================================
    -- Type definitions
    -- ======================================================================
    CREATE TYPE cust_address_typ
      OID '82A4AF6A4CD1656DE034080020E0EE3D'
      AS OBJECT
        ( street_address     VARCHAR2(40)
        , postal_code        VARCHAR2(10)
        , city               VARCHAR2(30)
        , state_province     VARCHAR2(10)
        , country_id         CHAR(2)
    REM ===========================================================================
    REM Create phone_list_typ varray to be varray column in customers table.
    REM ===========================================================================
    CREATE TYPE phone_list_typ
      OID '82A4AF6A4CD2656DE034080020E0EE3D'
      AS VARRAY(5) OF VARCHAR2(25);
    REM ===========================================================================
    REM Create customers table.
    REM The cust_geo_location column will become MDSYS.SDO_GEOMETRY (spatial)
    REM datatype when appropriate scripts and data are available.
    REM ===========================================================================
    DEFINE vscript = ?/demo/schema/order_entry/ccus_&vrs
    @&vscript
    CREATE UNIQUE INDEX customers_pk
       ON customers (customer_id) ;
    REM Both table and indexes are analyzed using the oe_analz.sql script.
    ALTER TABLE customers
    ADD ( CONSTRAINT customers_pk
          PRIMARY KEY (customer_id)
    REM ===========================================================================
    REM Create warehouses table;
    REM  includes spatial data column wh_geo_location and
    REM  XML type warehouse_spec (was bug b41)
    REM ===========================================================================
    DEFINE vscript = ?/demo/schema/order_entry/cwhs_&vrs
    @&vscript
    CREATE UNIQUE INDEX warehouses_pk
    ON warehouses (warehouse_id) ;
    ALTER TABLE warehouses
    ADD (CONSTRAINT warehouses_pk PRIMARY KEY (warehouse_id)
    REM ===========================================================================
    REM Create table order_items.
    REM ===========================================================================
    CREATE TABLE order_items
        ( order_id           NUMBER(12)
        , line_item_id       NUMBER(3)  NOT NULL
        , product_id         NUMBER(6)  NOT NULL
        , unit_price         NUMBER(8,2)
        , quantity           NUMBER(8)
    CREATE UNIQUE INDEX order_items_pk
    ON order_items (order_id, line_item_id) ;
    CREATE UNIQUE INDEX order_items_uk
    ON order_items (order_id, product_id) ;
    ALTER TABLE order_items
    ADD ( CONSTRAINT order_items_pk PRIMARY KEY (order_id, line_item_id)
    CREATE OR REPLACE TRIGGER insert_ord_line
      BEFORE INSERT ON order_items
      FOR EACH ROW
      DECLARE
        new_line number;
      BEGIN
        SELECT (NVL(MAX(line_item_id),0)+1) INTO new_line
          FROM order_items
          WHERE order_id = :new.order_id;
        :new.line_item_id := new_line;
      END;
    REM ===========================================================================
    REM Create table orders, which includes a TIMESTAMP column and a check
    REM constraint.
    REM ===========================================================================
    DEFINE vscript = ?/demo/schema/order_entry/cord_&vrs
    @&vscript
    CREATE UNIQUE INDEX order_pk
    ON orders (order_id) ;
    ALTER TABLE orders
    ADD ( CONSTRAINT order_pk
          PRIMARY KEY (order_id)
    REM ===========================================================================
    REM Create inventories table, which contains a concatenated primary key.
    REM ===========================================================================
    CREATE TABLE inventories
      ( product_id         NUMBER(6)
      , warehouse_id       NUMBER(3) CONSTRAINT inventory_warehouse_id_nn NOT NULL
      , quantity_on_hand   NUMBER(8)
    CONSTRAINT inventory_qoh_nn NOT NULL
      , CONSTRAINT inventory_pk PRIMARY KEY (product_id, warehouse_id)
    REM ===========================================================================
    REM Create table product_information, which contains an INTERVAL datatype and
    REM a CHECK ... IN constraint.
    REM ===========================================================================
    CREATE TABLE product_information
        ( product_id          NUMBER(6)
        , product_name        VARCHAR2(50)
        , product_description VARCHAR2(2000)
        , category_id         NUMBER(2)
        , weight_class        NUMBER(1)
        , warranty_period     INTERVAL YEAR TO MONTH
        , supplier_id         NUMBER(6)
        , product_status      VARCHAR2(20)
        , list_price          NUMBER(8,2)
        , min_price           NUMBER(8,2)
        , catalog_url         VARCHAR2(50)
        , CONSTRAINT          product_status_lov
                              CHECK (product_status in ('orderable'
                                                      ,'planned'
                                                      ,'under development'
                                                      ,'obsolete')
    ALTER TABLE product_information
    ADD ( CONSTRAINT product_information_pk PRIMARY KEY (product_id)
    REM ===========================================================================
    REM Create table product_descriptions, which contains NVARCHAR2 columns for
    REM NLS-language information.
    REM ===========================================================================
    CREATE TABLE product_descriptions
        ( product_id             NUMBER(6)
        , language_id            VARCHAR2(3)
        , translated_name        NVARCHAR2(50)
    CONSTRAINT translated_name_nn NOT NULL
        , translated_description NVARCHAR2(2000)
    CONSTRAINT translated_desc_nn NOT NULL
    CREATE UNIQUE INDEX prd_desc_pk
    ON product_descriptions(product_id,language_id) ;
    ALTER TABLE product_descriptions
    ADD ( CONSTRAINT product_descriptions_pk
         PRIMARY KEY (product_id, language_id));
    ALTER TABLE orders
    ADD ( CONSTRAINT orders_sales_rep_fk
          FOREIGN KEY (sales_rep_id)
          REFERENCES hr.employees(employee_id)
          ON DELETE SET NULL
    ALTER TABLE orders
    ADD ( CONSTRAINT orders_customer_id_fk
          FOREIGN KEY (customer_id)
          REFERENCES customers(customer_id)
          ON DELETE SET NULL
    ALTER TABLE warehouses
    ADD ( CONSTRAINT warehouses_location_fk
          FOREIGN KEY (location_id)
          REFERENCES hr.locations(location_id)
          ON DELETE SET NULL
    ALTER TABLE customers
    ADD ( CONSTRAINT customers_account_manager_fk
          FOREIGN KEY (account_mgr_id)
          REFERENCES hr.employees(employee_id)
          ON DELETE SET NULL
    ALTER TABLE inventories
    ADD ( CONSTRAINT inventories_warehouses_fk
          FOREIGN KEY (warehouse_id)
          REFERENCES warehouses (warehouse_id)
          ENABLE NOVALIDATE
    ALTER TABLE inventories
    ADD ( CONSTRAINT inventories_product_id_fk
          FOREIGN KEY (product_id)
          REFERENCES product_information (product_id)
    ALTER TABLE order_items
    ADD ( CONSTRAINT order_items_order_id_fk
          FOREIGN KEY (order_id)
          REFERENCES orders(order_id)
          ON DELETE CASCADE
    enable novalidate
    ALTER TABLE order_items
    ADD ( CONSTRAINT order_items_product_id_fk
          FOREIGN KEY (product_id)
          REFERENCES product_information(product_id)
    ALTER TABLE product_descriptions
    ADD ( CONSTRAINT pd_product_id_fk
          FOREIGN KEY (product_id)
          REFERENCES product_information(product_id)
    REM ===========================================================================
    REM Create cross-schema synonyms
    REM ===========================================================================
    CREATE SYNONYM countries FOR hr.countries;
    CREATE SYNONYM locations FOR hr.locations;
    CREATE SYNONYM departments FOR hr.departments;
    CREATE SYNONYM jobs FOR hr.jobs;
    CREATE SYNONYM employees FOR hr.employees;
    CREATE SYNONYM job_history FOR hr.job_history;
    REM ===========================================================================
    REM Create sequences
    REM ===========================================================================
    CREATE SEQUENCE orders_seq
    START WITH     1000
    INCREMENT BY   1
    NOCACHE
    NOCYCLE;
    REM ===========================================================================
    REM Need commit for PO
    REM ===========================================================================
    COMMIT;And yes, you must install at least Oracle SE to get completely this sample schemas on your file system :(

  • Sampling scheme

    Gurus,
    I have created a sampling scheme & assigned it in sampling procedure.When i record results if the sample size is 15 if i have
    inspected only 12 & enter the value the system is giving an error msg,
    Diagnosis
    You can directly confirm the number of nonconforming units or the system calculates the number of nonconforming units from the number of sample units beyond the tolerance range.
    System Response
    The number of nonconforming units cannot be greater than the number of inspected test units.
    Procedure
    Check and correct your entry.
    What should be done to overcome this bcoz in real time all the sample might not be inspected

    Are you sure you had the numbers in the right colums?  I've found that people easily confuse and put the wrong values in the colums "No. inspected" and "No. defects". 
    Also, what was the scope you set for your characteristics?
    If you had fixed scope on the characteristics, the system is expecting 15 results and will throw errors if you don't record 15. 
    From your error message however, it looks like you have attempted to record a higher number of defective units then what was inspected.  So if you make 1000 widgets and in QM you are inspecting 15, you can't record more then 15 defective units. If you inspect more then 15, then you have to change the number of inspected units to the actual number of units you actually inspected.  Then you can confirm up to that many defective units.
    Craig

  • Display real time data on a plot in a sub VI and main VI

    I am building a program to measure and plot real time data. Program has several steps so I build few Sub VIs to make it simple. My problem is I am plotting real time data in my SUB VI(it works fine), but in my main program when I try to get the same plot its not real time data plot. Plots appear at the end of the program.  All tips and help would be greatly appreciated. Thanks
    I have attached my main VI and Sub VI
    Solved!
    Go to Solution.
    Attachments:
    main VI.jpg ‏116 KB
    Sub VI.jpg ‏242 KB

    LabVIEW program is based on DATA FLOW, now, until the subVI will finish execution (while loop will be stopped), the execution in Main VI will not proceed any further.
    If you can pass the reference of 'XY Graph' inside the subVI, you'll be able to update the plot (placed on MAIN VI) from within subVI.
    Find attached example for your reference.
    I am not allergic to Kudos, in fact I love Kudos.
     Make your LabVIEW experience more CONVENIENT.
    Attachments:
    Example [LV 90].zip ‏26 KB

  • Issue to make mandatory field(SELD/BBD and Date of manufactor)in migo for particular Material type and material Number starting with'1' series.

    Hi friend,
    i have issue regarding mandatory  self life field (SELD/BBD and Date of manufactor)in migo for Batch Tab for particular Material start with '1' Series and material type.
    i want to make mandatory above field during GR from migo..any one let me know exit or badi for that to full fill this goal..
    Regard's,
    shaikh Khalid.

    Hi Shaikh
    First of all Thread is not closed seconldy as a good practice if you have resolved your issue kindly document it here so that it may help someone in future
    Nabheet

  • How to count occurences of a certain string in incoming real-time data? Also displaying RTC data. Current VI included

    I use LabView student Express 7 on a Windows XP system.
    Time-frame: we are doing final integrations for our balloon experiment today. We just got told that the press wants to view real-time data, which we haven't programmed for. I need help to get a working VI at the latest by 25.02.2004 before 0800(morning) (GMT+1).
    Note on license
    It is a student balloon flight, and the data will not be used in scientific work, so the I am not breaking any license agreements (I hope).
    Problem synopsis:
    The balloon continually transmits data at 9600baud. The data is a semi-repeating header followed by a constant lenght data-package arranged like this:
    BEXUS[h][m][s]BEXUS[h][m][s]
    [Raw binary data, 7channels*8sub-channels*8bits]
    What the groundstation is doing right now:
    Take all incomming data and save (append) the data to a file without any data-handling. (We figured we would go post-processing).
    What I need to change in less than 24 hours:
    - Add a "package" counter
    - Add a display of the clock data (RTC)
    How I planned to implement the changes:
    -RTC display:
    The RTC data is in BCD format, since that means that if you look at the data as hex numbers, you get the hours and minutes and seconds out in "clear text". That is 12 hours is 0x12hex. I figured that I can do a match pattern BEXUS and pass the "after substring" to another match pattern BEXUS from which I feed the "before substring" to a type-cast VI (casting string to u8) and displaying that, which should give me a display of "123000" for the time 12:30:00... I couldn't get it to work at all when I tried out the supplied "beta" vi.
    - Package counter:
    Counting how many BEXUS that gets detected and dividing by 2. I don't know how to do this. I've looked on the forum (a good thread on the problem: "how do I count the number of *'s in a string") but these use either loops or arrays... and I'm not sure how this works when I'm getting the data in at realtime. I cant make an array and then count it, since then the array would grow fast and possibly interfere with saving of the data??? Saving the data is critical.. without that file we cant do post-processing.
    Since my time is so limited (I'm not even supposed to do the groundstation software but they called on me in the last minute because no-one else had time/wanted too/could do it) I hope that you could make an exception and provide me with working VI's (based on the one I have attached) so that I can show something to the press! (Free comercial for NI!! Since the student version shows the National Instruments water-mark on all VI's!!! Possible TV time!!)
    Thanks!
    PS: even if you are to late (after 25) post anyway!
    Why:
    -I can learn from it
    -the launch might be delayed due to weather conditions
    -others might find it amusing!
    Thanks again!
    Attachments:
    BexusII_groundstation.vi ‏46 KB

    I have a valid example data file attached to this thread.
    If you open BEXTEST.bin in a hex-editor of your choice, you'll see the BEXUS as 42 45 58 55 53 and then the time as 00 28 09 etc.
    I couldn't get Joe Guo's VI to work. It doesn't count packages correctly, and the time is not displayed correctly either.
    The file was saved using a straight save to file VI.
    The data is from actual launching area tests performed a few mintues ago. The time displayed is "On time" e.g. how long the gondola has been powered up.
    I have a spare T-junction, so I can hook into the balloon real-time data as we fly, in case anyone care to see if they can figure out why the latest version of Joe Guo's program is not displaying correctly.
    I will monitor this
    thread during and after flight to see if anyone can make it in time!
    Thanks for the great effort!!
    Attachments:
    bextest.bin ‏53 KB

  • XML Schema file in data definition

    Hi,
    I am creating a custom RTF template for a seeded data definition in the eBusiness Suite. The data definition has an XML Schema file loaded however this does not include all the data required for the template. Running the report to just generate the XML (i.e. no template) I can see that the additional data is available at this level just not included in the XSD.
    I have built the template including the additional elements from the XML and expected to have to create a custom data definition as well to add the additional elements to the XSD. However the template is working against the seeded definition and successfully picking up the elements that are not included in the XSD.
    Has anybody else come across this situation? And are there any potential issues? I could create a custom data definition as well to load an amended XSD but this would be a more invasive change as it would involve end dating the seeded data definition which I would prefer not to do. The report is not being generated from a concurrent request so I cannot just create a custom request and leave the seeded report untouched.
    Thanks,
    Sarah

    The XML Schema is only required if you are using a PDF template (the template file itself is PDF) with field-mapping. For other template types the schema is optional.
    The XML Schema is used to define a "class" of XML documents, so to speak. The define the semantics and structure of the XML documents more thoroughly.
    If you have closely related reports, you could use a single schema to standardize (in a sense) the XML to be output from the class of related reports. In that case you could have a single schema file uploaded to multiple reports/data-definitions.
    The W3C has good documentation on XML Schema.
    If you want to see samples, you can use tools such as dtd2xs to create sample schemas from your DTD.

  • HP in Real Life - Tell us your story.

    Hi everyone,
    In support of the HP in Real Life program, I thought we would support it by not only talking about it, but also starting some stories here about how we use our HP products to make our life, and the lives of those around us, better. 
    Tell us how HP products help you tackle everyday tasks like keeping family connected across the country or around the globe, balancing work and play, or just planning life and events. 
    You can either share your experience here, or go over to the Real Life website to share your story in videos, pictures or words to show others what HP helps you accomplish. If your story is selected, you'll be featured on the HP in Real Life website! 
    What started as a thoughtful gift from a friend has quickly turned into the thriving hub of business for Rebecca, the owner and founder of Danette Kenneth Photography. Learn how she uses her Compaq Presario CQ50 Notebook PC to help run and promote her business.
    Andy and Irwin have been passionate about ham radio for years. Learn how they use the latest technology in their HP and Compaq laptops to establish connections, translate Morse code and log new contacts from around the world.
    Have a great week. See you on the Boards.
    GarenT
    HP Product Expert for the Officejet Pro X Series.
    Was your question answered? Mark it as an Accepted Solution!
    See a great post? Give it a Kudos!

    Hello,
    There are many posts on this site concerning issues such as this from your region. Many who have posted have reported that, when they escalated through their proper support channels, they have been told that the problem is network congestion. But, "blame" shifts back and forth between RIM and the local mobile service providers. Consequently, I recommend that all who are experiencing this problem contact their mobile service providers and open a formal support ticket on this issue. The more of those that get placed, the more likely they will actually do something about it.
    Good luck!
    Occam's Razor nearly always applies when troubleshooting technology issues!
    If anyone has been helpful to you, please show your appreciation by clicking the button inside of their post. Please click here and read, along with the threads to which it links, for helpful information to guide you as you proceed. I always recommend that you treat your BlackBerry like any other computing device, including using a regular backup schedule...click here for an article with instructions.
    Join our BBM Channels
    BSCF General Channel
    PIN: C0001B7B4   Display/Scan Bar Code
    Knowledge Base Updates
    PIN: C0005A9AA   Display/Scan Bar Code

  • Oracle 9i Sample Schema running in client/server mode?

    I've got Oracle 9i up on RH7.3 and have successfully executed the sample schema on the Server using the mksample.sql script.
    Now the challenge is to run these same (sample schema) SQL scripts across a network using a 9i client. I copied the demo/schema files from the server installation to the client but when I invoke the mksample script the passwords have to include a connect_identifier which, in this environment, causes the subordinate scripts to get confused.
    Has anyone else attempted to do this?

    You did not have to change any of your scripts to put the connect_identifier.
    Within SQL*Plus you can use the "SET INSTANCE <connect_identifier>" command to make that Oracle instance
    your default for this session. After you do this, all your connect statements will not require a
    connect_identifier.
    ==============================================================
    $ sqlplus /nolog
    SQL> set instance ORA901
    Oracle9i Enterprise Edition Release 9.0.1.1.1 - Production
    With the Partitioning option
    JServer Release 9.0.1.1.1 - Production
    SQL> connect scott/tiger
    Connected.
    SQL>
    ==============================================================
    As you can see, after the SET INSTANCE command, the connect command does not need a connect_identifier.

Maybe you are looking for

  • External HDD to use with logic/sample libraries

    So my current external HDD just freaked out, and after a year of using it with my MBP, it decided that its format was NTFS. Not sure how that works, but I dont want to have to experience that again . Anyway, I'm buying a new one, and was considering

  • Display the records in forms when creating in a database

    hi to all i have a problem in forms, run in c\s using form6i.. How can i display the records when i create a records to database and automatically display in forms when i open the forms or when creating a records? example: in form#1 , this is where i

  • ..........Error unable to Update

    Hi all I have table which contains the request data. requested by User. There are 3 admins each at different region which can take action on the requests send by thier regional teams/users. For Region A , it works fine that the Admin opens the Form,

  • Bundle Size Listing

    Is there a way to list the size (in bytes or whatever) of all of the Bundles in a content-repo? We are running low on space and trying to determine which bundles (ours or ZPM) we can remove to restore space. Thanks.

  • Issue with IISReset in MOSS server

    Hello, We have a Moss 2007 Server with Windows 2003 Server installed. When we do an IISReset on one of teh WFE server, we get the below error in event log. IIS Reset encountered an error while stopping services, which was requested by.The logged data