ORA-00933 on table alias introduced with the AS keyword

Oracle 10g is unable to parse the following query:
SELECT ALL "COUNTRY"."C_NAME" AS "NAME" FROM "C_COUNTRY" AS "COUNTRY";
However, the following query parses and executes all right:
SELECT ALL "COUNTRY"."C_NAME" AS "NAME" FROM "C_COUNTRY" "COUNTRY";
It seems that a correlation name (a table alias) cannot be preceded by the AS keyword. IIRC the SQL standard allows this so I am very surprised.

I stand corrected. Ingres does support "AS" to alias the table name:
* select ALL "S"."CODE" FROM "SECURITY_CODES" AS "S"\g
Executing . . .
┌─────────────┐
│code                 │
├─────────────┤
│                       0│
│                       1│
│                       2│
│                       3│
│                       4│
│                       5│
│                       6│
│                       7│
│                       8│
│                       9│
└─────────────┘
(10 rows)
continue
*Not a syntax I've ever used though.

Similar Messages

  • Essbase answers - None of the fact tables are compatible with the query request "member"

    Hi,
    I have modelled an Essbase database into the repository.
    If I pull the measure, period and year dimension in and filter on the year (member) and display the year (member) along with the period (alias) and measure it errors with =>
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 43113] Message returned from OBIS. [nQSError: 43119] Query Failed: [nQSError: 14020] None of the fact tables are compatible with the query request Fiscal Year.Fiscal Year Code. (HY000)
    However, all other things being equal if I change the year displayed to the alias then it works.
    Anyone tell me why??
    Is there a limitation that Essbase brings through that you cannot view what you filter on?
    thanks,
    Robert.

    Hi
    i have done the content level setting in each of the table, D1,F1 and F2(LTS), now i am getting the following error..
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 15018] Incorrectly defined logical table source (for fact table Gl Sets Of Books) does not contain mapping for [Code Combinations.Code Combinations.Affiliate, GL Balances.GL Balances.Currency Code, GL Balances.GL Balances.PTD_Balance, Gl Sets Of Books.Gl Sets Of Books .SoB Name]. (HY000)
    Gl Balances : D1
    Code Commbination: F1
    Gl Sets Of Books : F2
    I have checked the joins in physical and BMM layer..all are fine..

  • Content Tab: None of the fact tables are compatible with the query request

    Hi All,
    **One thing I am not clear yet of all my years with OBIEE is working with the content tab in BMM.**
    I have made a rpd the joins in physical layer as shown below:
    https://picasaweb.google.com/114804305606242416264/OBIEEError#5663056545119428530
    And the BMM layer as:
    https://picasaweb.google.com/114804305606242416264/OBIEEError#5663056519553812930
    Error I am getting when i run a request from the 3 columns from the selected 3 tables is:
    Dim - Comment Code Details
    Fact - Complaint
    Dim - Service Details
    Error Codes: OPR4ONWY:U9IM8TAC:OI2DL65P
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 14020] None of the fact tables are compatible with the query request Sr Num:[DAggr(Fact - Complaint.Sr Num by [ Dim - Service Details.Sr Cat Type Cd, Dim - Comment Code Details.Cmtcode name] )]. (HY000).
    I get no error for consistency.. I read everywhere and I know i need to set the appropriate aggregation levels in the various dims and facts LTS properties to help OBIEE understanding our model, but how to do that.. how do i decide... how should I approach, what should be the aggregation level, what details.
    When i click More button i see different options: Copy, Copy From, Get Levels, Check Level, what do these mean.
    Aggregation Content, group by - Logical Level or Column which one should i choose and how should I decide.
    Can anyone explain the Content Tab in details and from scratch with some example and why we get these errors.... I know many people who are well versed with many other things related to RPD but this. A little efforts of explaining from you guys will really be appreciated.
    Thanks in advance,
    Dev

    Hi Deepak,
    Option 1:
    My tables in physical layer are joined as below:
    D1--> F1 <--D2--> F2 <--D3
    Same way i model it in BMM
    D1--> F1 <-- D2--> F2 <--D3
    Here D1 is non Conformed Dimension for F2 and D3 is non Conformed dim for F1. Later create Dimensional hierarchies, I tried setting up the content levels
    I go Sources>content tab of Fact F1 I set
    Dimensions----------- Logical level
    D1---------------------- D1 Detail
    D2---------------------- D2 Detail
    D3---------------------- D3 Total
    then, I go Sources>content tab of Fact F2 I set
    Dimensions----------- Logical level
    D1---------------------- D1 Total
    D2---------------------- D2 Detail
    D3---------------------- D3 Detail
    Then, I also go in all the dimensions and set their content levels to Details, but it still gives me errors not sure where I am going wrong in setting the content levels.
    I need to know whether the way I have modeled it in BMM is right,
    Option 2:
    I can combine the two facts in a single Logical Fact or the above design should also work.
    (F1&F2)<--D1, D2 , D3 joined separately using complex logical joins.
    what will be the content tab details?
    Thanks,
    Dev

  • Fact tables are compatible with the query request

    Hi,
    i am using 11g.In 10g working fine without any error.After migrate 10g into 11g below error will returning. What is the problem and How we will fix.Could you pls let me know.Thanks
    Error Codes: OPR4ONWY:U9IM8TAC:OI2DL65P
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 43113] Message returned from OBIS. [nQSError: 43119] Query Failed: [nQSError: 14020] None of the fact tables are compatible with the query request FACT_AGENT_TIME_STATISTICS.FKEY. (HY000)
    SQL Issued: SELECT s_0, s_1, s_2, s_3, s_4, s_5, s_6, s_7, s_8, s_9, s_10 FROM ( SELECT 0 s_0, "Team Performance"."DIMENSION - LOCATION"."COUNTRY CODE" s_1, "Team Performance"."DIMENSION - TIME"."BUSINESS DATE" s_2, CASE WHEN "Team Performance"."DIMENSION - TEAM"."ROLE" ='TEAM LEADER' THEN 'Team Leader' ELSE "Team Performance"."DIMENSION - TEAM"."TEAM" END s_3, CASE WHEN '30 Mins Interval' ='15 Mins Interval' THEN "Team Performance"."DIMENSION - TIME"."15 Mins Interval" ELSE "Team Performance"."DIMENSION - TIME"."30 Mins Interval" END s_4, ((SUM("Team Performance"."FACT - AGENT CALL STATISTICS"."TIME - ACD CALL HANDLING"+"Team Performance"."FACT - AGENT CALL STATISTICS"."TIME - AFTER CALL WORK (ACW)")+SUM("Team Performance"."FACT - AGENT TIME STATISTICS"."AVAILABLE TIME"))/60)/(COUNT(DISTINCT "Team Performance"."FACT - AGENT TIME STATISTICS"."DATE ID")*"Team Performance"."FACT - AGENT TIME STATISTICS"."INTERVAL TYPE") s_5, (SUM("Team Performance"."FACT - AGENT TIME STATISTICS"."AUX3 - TRAINING"+"Team Performance"."FACT - AGENT TIME STATISTICS"."AUX4 - MEETING"+"Team Performance"."FACT - AGENT TIME STATISTICS"."AUX5 - PROJECT"+"Team Performance"."FACT - AGENT TIME STATISTICS"."AUX6 - COACHING")/60)/(COUNT(DISTINCT "Team Performance"."FACT - AGENT TIME STATISTICS"."DATE ID")*"Team Performance"."FACT - AGENT TIME STATISTICS"."INTERVAL TYPE") s_6, (SUM("Team Performance"."FACT - AGENT TIME STATISTICS"."STAFFED TIME")/60)/(COUNT(DISTINCT "Team Performance"."FACT - AGENT TIME STATISTICS"."DATE ID")*"Team Performance"."FACT - AGENT TIME STATISTICS"."INTERVAL TYPE") s_7, COUNT(DISTINCT "Team Performance"."FACT - AGENT TIME STATISTICS"."DATE ID")*"Team Performance"."FACT - AGENT TIME STATISTICS"."INTERVAL TYPE" s_8, MIN("Team Performance"."DIMENSION - TIME"."FULL DATE TIME") s_9, REPORT_AGGREGATE(((SUM("Team Performance"."FACT - AGENT CALL STATISTICS"."TIME - ACD CALL HANDLING"+"Team Performance"."FACT - AGENT CALL STATISTICS"."TIME - AFTER CALL WORK (ACW)")+SUM("Team Performance"."FACT - AGENT TIME STATISTICS"."AVAILABLE TIME"))/60)/(COUNT(DISTINCT "Team Performance"."FACT - AGENT TIME STATISTICS"."DATE ID")*"Team Performance"."FACT - AGENT TIME STATISTICS"."INTERVAL TYPE") BY CASE WHEN '30 Mins Interval' ='15 Mins Interval' THEN "Team Performance"."DIMENSION - TIME"."15 Mins Interval" ELSE "Team Performance"."DIMENSION - TIME"."30 Mins Interval" END, CASE WHEN "Team Performance"."DIMENSION - TEAM"."ROLE" ='TEAM LEADER' THEN 'Team Leader' ELSE "Team Performance"."DIMENSION - TEAM"."TEAM" END) s_10 FROM "Team Performance" WHERE (("DIMENSION - TIME"."BUSINESS DATE" BETWEEN timestamp '2012-11-23 00:00:00' AND timestamp '2012-11-23 00:00:00') AND ("DIMENSION - LOCATION"."COUNTRY CODE" = 'US') AND ("DIMENSION - LOCATION".DEPARTMENT = 'CSG')) ) djm FETCH FIRST 65001 ROWS ONLY

    Back up your repository before following these steps:
    1 Reduce the problem report to minimum number of columns required to generate the error. This will usually identify which dimension and which fact are incompatible.
    2 Open the repository in the Administration tool and verify that the dimension and fact table join at the physical layer of the repository
    3 Verify that there is a complex join between the dimension and the fact in the business layer.
    4 Check the logical table sources for the fact table. At least one of them must have the Content tab set to a level in the hierarchy that represents the problem dimension. This is usually the detailed level.
    5 Check the logical table source Content tab for the dimension table. Unless there is a valid reason, this should be set to blank.
    6. Save any changes to the repository.

  • Error: None of the fact tables are compatible with the query request

    Hi experts,
    I have one confirm dimension D1 and other two fact table F1 and F2 (F1 and F2 are connected to D1)
    when i create a report from D1 and F1 the report is running fine. But pull the column from F2 also in this report i am getting the following error
    None of the fact tables are compatible with the query request Code Combinations
    please suggest on the same.
    Regards,
    S

    Hi
    i have done the content level setting in each of the table, D1,F1 and F2(LTS), now i am getting the following error..
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 15018] Incorrectly defined logical table source (for fact table Gl Sets Of Books) does not contain mapping for [Code Combinations.Code Combinations.Affiliate, GL Balances.GL Balances.Currency Code, GL Balances.GL Balances.PTD_Balance, Gl Sets Of Books.Gl Sets Of Books .SoB Name]. (HY000)
    Gl Balances : D1
    Code Commbination: F1
    Gl Sets Of Books : F2
    I have checked the joins in physical and BMM layer..all are fine..

  • None of the dimension tables are compatible with the query request

    Hi,
    i am experiencing the below error while querying columns alone from employee dimension (w_employee_d) in workforce profile SA. There is only one column in my report which is employee number coming from employee dimension. when i query other information like job, region, location etc i am not getting any error. the below error appears only when querying columns from employee dimension. the content tab level for the LTS of employee dimension is set to employee detail.
         View Display Error
    Odbc driver returned an error (SQLExecDirectW).
    Error Details
    Error Codes: OPR4ONWY:U9IM8TAC:OI2DL65P
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 43113] Message returned from OBIS. [nQSError: 43119] Query Failed: [nQSError: 14077] None of the dimension tables are compatible with the query request Dim - Employee.Employee Number. (HY000)
    SQL Issued: SELECT 0 s_0, "Human Resources - Workforce Profile"."Employee Attributes"."Employee Number" s_1 FROM "Human Resources - Workforce Profile" FETCH FIRST 65001 ROWS ONLY.
    couldn't able to know the exact reason. Any suggestions would be highly appreciated.
    Regards.

    hi user582149,
    It is difficult to answer you question with such a little amount of details. Could you specify:
    - how many facts/dimensions are you using in the query?
    - what is the structure of your Business Model?
    - which version of OBI are you using?
    - what does your log say?
    I hope to tell you more having the information above
    Cheers

  • "None of the fact tables are compatible with the query request " error

    I've got a situation where I have two facts(Fact_1, Fact_2) and three dimensions(dim_1,dim_2,dim_3) in 1 subject area. I've got dimension hierarchies setup for all the dimension tables.
    Dim_1 is one to many to Fact_1
    Dim_2 is one to many to Fact_2
    Dim_3 is one to many to both Fact_1 and Fact_2
    I've set up the content levels for the LTS for the Facts so that they are the lowest grain for dimensions they join to and the grand total grain for dimensions they do not join to.
    My rpd is consistent. When I run a report using an attribute from Dim_3 and Dim_1 or Dim_3 and Dim_2, the report comes back fine.
    But if I try to run a report using all three Dim tables, I get an error and the message "None of the fact tables are compatible with the query request ".
    First of all, is it possible to make a report using all three dimensions?
    Second, what's the best way to trouble shoot this error? Why are none of the fact tables compatible? I thought as long as the aggregation levels were set to grand total for non-shared dimensions, Answers would be able to create the report properly.
    Any advise would be greatly appreciated.
    Thanks!
    -Joe

    OBIEE is looking for a fact that can link ALL the dimensions together. This is also known as the implicit fact ... you don't have a fact that can relate all the dimensions - you have 2 facts that together they can. Perhaps you need to great a single logical fact that has both LTS for your physical facts and try it that way.
    Then you'd have Dim1, Dim 2, Dim3 all being able to join to Fact1 (which is made of physical facts 1 & 2).

  • [svn:fx-trunk] 9343: Tweak for minor scanner issue introduced with the DesignLayer change.

    Revision: 9343
    Author:   [email protected]
    Date:     2009-08-17 07:15:39 -0700 (Mon, 17 Aug 2009)
    Log Message:
    Tweak for minor scanner issue introduced with the DesignLayer change.
    QE notes: N/A
    Doc notes: N/A
    Bugs: SDK-22714
    Reviewer: Gaurav
    Tests run: Compiler cyclone.
    Is noteworthy for integration: No.
    Ticket Links:
        http://bugs.adobe.com/jira/browse/SDK-22714
    Modified Paths:
        flex/sdk/trunk/modules/compiler/src/java/flex2/compiler/mxml/dom/MxmlScanner.java

  • How Do I Sort a Group of Photos on the Same Page with the same Keyword Tag?

    I want to group similar photos on the same page with the same keyword tag.  I have tried dragging and dropping without any success.  Can anyone help?

      Create an album, then search using a keyword. Select all the photos produced by the search and drag them into the album.
    You can drag photos into any order in Albums and they remain in your chosen order.
     

  • How to update fields in the target table in correspondance with the source file values

    Environment: win7, SQL server 2008 R2
    Application: Microsoft Management SQL Studio 2008 R2, Business Intelligence 2008 - SSIS
    SSIS competency level: Novice
    Problem: I have been trying to update some of the fields in the destination table,student table, in reference to data set in the staging table and ssn table.  I was able to insert/load new data to the destination using look up transformation
    while the driver is ssn (data mapping) but i couldn't know how to update some of the fields in the student table while keeping the orignal pn_id of both tables(ssn and student tables), because pn_id already exists in the SSN table and student table. There
    are other records also associated with the pn_id so I am not allowed to update the pn_id in the destination tables. For example,
    SSN Table (pn_id,ssn)
    ('000616850',288258466)
    ('002160790',176268917)
    Staging Table (ssn, id, pn_id, name, subject, academic year, comments)
    (288258466, 1001, '770616858',Sally Johnson, English,A, 2005,'great student')
    (176268917, 1002, '192160792',Will Smith, Math,38000,C, 2014,'no comments')
    (444718562, 1003, '260518681',Mike Lira, Math,38000,B, 2013,'no comments')
    Student Table (destination table)(id,pn_id,subject,academic year, grade, comments):
    (1001, '000616850', ' ',' ', ,'')
    (1002, '002160790', ' ',' ', ,'')
    Expected Results:
    My goal is to have student table updated as the following:
    Student Table
    (1001, '000616850', 'English','A' ,2005 ,'great student')
    (1002, '002160790', 'Math ',' C',2014 ,'no comments')
    please advise

    Why can't you use simple UPDATE command in EXECUTE SQL Task as below,
    DROP TABLE SSN
    DROP TABLE STAGING
    DROP TABLE STUDENT
    CREATE TABLE SSN(pn_id VARCHAR(100),ssn BIGINT)
    INSERT INTO SSN VALUES('000616850',288258466)
    INSERT INTO SSN VALUES('002160790',176268917)
    CREATE TABLE Staging (ssn BIGINT, id INT, pn_id BIGINT, name VARCHAR(100), subject VARCHAR(100),grade VARCHAR(10), [academic year] INT, comments VARCHAR(100))
    INSERT INTO Staging VALUES(288258466, 1001, '770616858','Sally Johnson', 'English','A', 2005,'great student')
    INSERT INTO Staging VALUES(176268917, 1002, '192160792','Will Smith', 'Math','C', 2014,'no comments')
    INSERT INTO Staging VALUES(444718562, 1003, '260518681','Mike Lira', 'Math','B', 2013,'no comments')
    CREATE TABLE Student(id INT,pn_id BIGINT,subject VARCHAR(100), [academic year] INT, grade VARCHAR(10), comments VARCHAR(100) )
    INSERT INTO Student VALUES(1001, '000616850', NULL,NULL,NULL ,NULL)
    INSERT INTO Student VALUES(1002, '002160790', NULL,NULL,NULL ,NULL)
    UPDATE Student SET Subject = C.Subject, [academic year]=C.[academic year], grade=C.grade,comments=C.comments
    FROM SSN A INNER JOIN Student B
    ON A.pn_id=B.pn_id INNER JOIN Staging C
    ON A.ssn = C.ssn
    SELECT * FROM Student
    Regards, RSingh

  • Database table not updating with the changes

    Hi friends,
                    Iam having a tablecontrol where iam gettting the data from the ztable for a particular key field.  When i change certain fields in the table control and when i click save, it should update the changes in the ztable.
       module save output.
       modify zslip_po1 from table x_po1_slip.
    here slpno is the key field.  if I change a field for a particular slpno it should update the changes for that slpno value only.
    but it is not updating the changes. although iam setting the slpno field dynamically.
    Please give me ideas.
    thanks and regards
    Murali Krishna T

    Hi,
    1. Have 2 Internal Tables, the first one must contain the original data(before change ) and the second one  contains the update data(after changed ie. table control data).
    2. loop the first table and read the data from the second table using index.
    3. inside the loop, compare the work area of both the internal table.
       if both are same, then no change was made to that particular record, so no need for updation
       if both are not same, then some changes were made on that record, so need to update.
    4. use modify command to update that particular record from the work area.
    i think in performance wise also this logic will be helpful, since we are updating only the changed record.
    since this is addon table, so i hope the no of records will also less, so no need to worry about loop.
    check with this, it may work out......
    thank you,
    SenthilPandi C
    Edited by: SenthilPandi ChandraSekaran on Feb 4, 2010 5:36 PM

  • Where-clause with the in-keyword. Query optimising.

    Hi I have following question.
    I have a query, e. g.
    select x, y from tableA where y in (a_1, a_2, ..., a_n)
    I have test it with n=1...5
    1. 5-6 ms
    2. 5-6 ms
    3. 150 - 200 ms
    4. 150 - 200 ms
    5. 180 - 250 ms
    There is a gap between n = 2 and n = 3. According to the execution plan the Oracle make a full table access if n >= 3. If n < 3 a an index is used. Is it possible to ensure that the index is always used?
    Additionally I have test the equivalent query:
    select x, y from tableA where y = a_1 or y = a_2 ... < = a_n
    It showsthe same behaviour.

    Hi,
    I wouldn't say that the optimizer is wrong here. I would look on some more imputs.
    How many values of y do you have in the table?
    please post the result of:
    select count(*),y
    from tablea
    group by y;
    to see how many values of each y-value exists in the table.
    For example, if there are only values 1..5 for y in the table a full table scan is the fastet answer for your query, because all values have to read!
    Regards
    Udo

  • Export aperture originals with the edits,keywords etc to external hard drive?

    Hi,
      I have filled more than half of my hard drive with 29,000 NEF files (mostly family photos).  I want to export them to a 2 TB external hard drive and then continue to import all new raw files in Aperture onto that same external hard drive.  I'm not sure how to do this?  I have all of the 29,000 photos back-ed up in an aperture vault on a 750 GB external drive so I feel relatively safe.  In other words, the originals and versions currently exist on my iMac internal hard drive and are backed up in an aperture vault on a portable hard drive; however, I want to have all my photos (with the existing edits) migrated/exported to an external hard drive to free up space on my iMac. Do I create a new library on an external disk?  If so, can I do that within Aperture using "Switch to new library" under File? Also, what is the safest way to get my originals and edits into that library. Is there a better plan than mine? thanks, Mike

    Hi Mike,
    I think Terence may have read your question too quickly.  I know that some definition of terms will help you express your goal, and in turn help us show you how to attain it.  I recommend _not_ proceeding with Terence's recommendation until it is clear what you want to do.
    "Photo" has no fixed or usable meaning in Aperture.  You camera records data every time you press the shutter release.  The recorded data is processed and saved in a file.  That file is almost always either a RAW file, or a JPEG.  RAW described a family of proprietary sensor-data formats.  There is no standard RAW format: every camera manufacturer uses their own formats, almost always varying them significantly from one camera to the next.  JPEG is a standard image-file format.  Computers can display image-format files.  Computers cannot display RAW files — it's just sensor data until the data is _converted_.  RAW converters (software) convert the RAW sensor data into a displayable image-format file.  OS X includes a RAW converter.  Aperture uses the OS X RAW converter.
    Aperture can import RAW files, and image-format files (and movies).  Once imported, the imported file is called an Original.  Importing a file creates a record in Aperture's database, and also creates a text-format file of instructions telling Aperture how to modify the imported file in order to create the Image.  Images are what you see in the Browser and in the Viewer.  The initial text-format file is null — the instructions say, in essence, "don't do anything to the Original, just display it (after converting it if needed)".  This text file is called a Version.  (Aperture also calls, wrongly and confusingly, imho, the Image you see a Version.)
    Aperture is a database.  Each record in the database points to a unique Version (a text file), and an Original.  There can be many Versions (and Images) for any one Original.
    You can store your Originals wherever you want.  Aperture makes it easy to relocate them, or store them inside your Library, any time you want.  They do not have all be in the same place.  The drives they are on don't even have to be on-line for you to use many functions in Aperture (if your Originals are off-line, you cannot adjust, export, or print Images).
    Your Versions (the text files) must be stored in your Library.  You cannot move them.
    Your storage strategy thus focuses on two things:
    - where to store your Aperture Library, and
    - where to store each Image's Original.
    Many users run out of storage space on their system drive at some point.  The commonly-recommended response is to re-locate some of your Originals to an external (or non-system) drive.  Alternatively, you might decide to move your Library to an external drive.  Or both.
    What you do depends on your needs.  With the information you've given, I recommend that you relocate your Originals to your dedicated external drive, and leave your Library on your system drive.  Select the Images whose Originals you with to relocate, and run the Aperture command "File ▹ Relocate Originals ... ", and select your external drive as the destination.
    You might then continue to import as you've been doing, that is, storing your new Originals on your system drive.  When it starts to get full again, just select the Images with managed Originals ("managed" means "stored in the Aperture Library"), and again run the relocate command.  You can filter your Images by whether the Image's Original is Managed or Referenced ("referenced" means "stored outside the Aperture Library").  Use the Filter HUD Rule "File Status".
    Note that Aperture's Vault _does not back up Referenced Originals_!  You must back them up yourself, using software other than Aperture.
    HTH,
    —Kirby.

  • How can I include a MySQL table and records with the PUT of my Dreamweaver site?

    I am using .php to validate a user code against one of four codes that I am providing to potential clients.  They enter that code, along with a default user id and the appropriate page is then displayed for them.  Everything is working fine on the desktop.  I need to have this table and the records created on my remote server.  I tried saving the table in the Connections folder, but that didn't help.  Everytime I enter a code (valid or invalid) on the live site, I get the following message:  "Table 'boundsauctions_com.promocode' doesn't exist".  Is there someway to include the table and records in my Dreamweaver folders for the site? 
    Thanks for any help that you can offer.

    You can't.  Dreamweaver, or FTP for that matter has no connection to the MySQL database.  If you have created a MySQL database locally and now wish to move/copy it to your server you'll need to create a back up of your local database and then log into your database on the server and then restore that back up.
    An even easier way to do this, and if you are working with MySQL much at all  a good investement is to get the commercial version of Navicat MySQL - this is the best MySQL administrator I've found to date and is well worth the few bucks to buy it.  With Navicat you can connect to your local MySQL database and also create a connection to your remote database, then literally drag and drop your local database contents to copy them to your server. This function alone justifies buying Navicat. A huge time saver.   And no I do not work for or have any share or ties to Navicat.  I'd recommend something else if I knew of anything better, but so far I don't.
    Hope this helps. If you need any more details let me know.
    Lawrence Cramer - *Adobe Community Professional*
    http://www.Cartweaver.com
    Shopping Cart for Adobe Dreamweaver
    available in PHP, ColdFusion
    Stay updated - http://blog.cartweaver.com

  • Table or view with the request status of all requests in a cube

    Dear all,
    I search the table or view which is used by Cube / Manage -> InfoProvider Administration (Requests).
    Target is to get all status information of all requests started by one InfoPackage.
    I found several tables alone but no view. I have currently
    RSREQICODS, RSMDATASTATE, RSSELDONE.
    Best regards.

    I've found a solution for the problem:
    *    Select all requests for one Infopackage
          SELECT distinct RNR from rsseldone into table ta_done where LOGDPID = IPak.
          LOOP AT ta_done
            INTO wa_done.
            WRITE: / 'Request: ', wa_done-RNR.
    *       Select the maximum Request Number in the Cube
            SELECT SINGLE ICUBE RNSIDLAST from RSICCONT Into wa_cont where RNR = wa_done-RNR.
            WRITE: / 'Icube: ', wa_cont-ICUBE.
            WRITE: / 'Max: ', wa_cont-RNSIDLAST.
    *       Check the Maximum Request with status technical and quality ok
            SELECT TECHOK QUALOK from RSMDATASTATE into table ta_state where INFOCUBE = wa_cont-ICUBE.
            LOOP AT ta_state INTO wa_state.
                WRITE: / 'ID TECHOK: ', wa_state-TECHOK.
                WRITE: / 'ID QUALOK: ', wa_state-QUALOK.
                IF wa_state-TECHOK > l_maxid AND wa_state-QUALOK > l_maxokid AND wa_state-QUALOK = wa_state-TECHOK.
                   l_maxokid = wa_state-TECHOK.
                ENDIF.
            ENDLOOP.
            IF wa_cont-RNSIDLAST > l_maxokid OR wa_cont-RNSIDLAST = 0.
              WRITE: / 'Request is running with Request: ', wa_cont-RNSIDLAST.
            ELSE.
              WRITE: / 'all Green'.
              l_cont = 0.
            ENDIF.
         ENDLOOP.

Maybe you are looking for

  • Can I send photos from my ipod to an android phone via bluetooth?

    If I can, how, because I have some photos that I want to send, but no one has an iphone, or ipod, so i need to send it to an android phone. Evertime i try bluetooth, my ipod doen't pick up the other signal, and keeps forgetting about devices. help?

  • I am getting this Error while trying to open the Open Document Syntax

    SYNTAX ="<a href=\"http://<MyServerIP>:8080/BUSINESSOBJECTS/ENTERPRISE115/INFOVIEW/OpenDocument/opendoc/openDocument.jsp?sDocType=wid&sDocName=59813204&sWindow=New&sRefresh=Y&lsSParameter="URLEncode(""[Parameter])"\" title=\"\" target=\"_self\" nav=\

  • Deleted files

    a number of files have been deleted from my pc. My ipod is set to auto update, is there any way to stop the files being deleted from there next time i connect it? Files seem to be disappearing at random. I haven't made any changes other than update t

  • Can't re-install snowleopard operating system

    Hi! Thanks for reading my question and helping me to solve my issue. I have an old macbook pro that was purchased around 2010, and I've been trying to re-install OS snowleopard (the original operating system) before selling this laptop. However, the

  • Maintain terms of Payment

    Hello to everyone, I have got a small problem and i hope someone help me!!! When i am trying to create a Payment, like that:  Payment day 5 to 2 mounths -day limit 30 -fixed day 5 -additional months 2 -Document day In my document the Posting date and