Working With Data Models !!!!

Hi,
I am working with TableViews can some one tell me about some good material I can go through for Data Models and how i can work with them.

Hi Emmanuel,
many people have given you some points for your help, but you didn't vice versa except once.
I will replace this text with the answer to your question (as far as I can) if you reward points (or at least reply why the offered help didn't work; be more responsive, please) on the following threads
How to create a new System !!!!!!!
How to use Visual Administrator !!!!
How to use Logger !!!!
Console Output !!!!  Need Help Badly !!!!
How to use onRowSelection !!!!
Best regards
Detlev

Similar Messages

  • Work with data from a table of another SAP R/3 system

    Hi,
    I want to work with data from a table of another system. I know there's a function called 'RFC_READ_TABLE' but I don't know really how it works. What I want to do is to get data from that system and pass it to the one where I'm working in order I can continue working with this data in the report.
    Besides, I'd like to select only the records that have some conditions.
    I'd appreciate if someone could tell how to do it, whether with FM 'RFC_READ_TABLE' or in another way.
    Thanks in advance,
    Gerard
    P.S: The systems are already connected in SM59

    Yes, you can use RFC_read_table .You can pass the number of columns but I recommend you to  bring all results in your internal table after that you can filter the records. Most of the time I have notice when you make RFC call with this FM to another system it works but it may also happen that other system controlling authorization and then it might give you nothing. As I face this problem on my last Project.
    Hope this’ll give you idea!!
    <b>P.S award the points.</b>
    Good luck
    Thanks
    Saquib Khan
    "Some are wise and some are otherwise"

  • SQL Developer with Data Modeler, No Design Menu Option

    Using SQL Developer 3.0.04 Build MAIN-04.34 with Data Modeler. I have created a logical design and wish to forward engineer this to a relational model.
    I cannot seem to find the Design Menu option to use the >> Engineer to Relational Model option which is available under the Design Menu on the Stand Alone version of SQL Developer Data Modeler.
    Have checked the following locations:
    Tools -> Data Modeler
    View -> Data Modeler
    File -> Data Modeler
    Right Click on my Logical Design
    I have ensure that all my entities have the Engineer To property set to a valid Relational Model.
    Is this a bug or am I missing a menu option / configuration setting?
    Thanks in advance for any help
    John

    Hi John,
    you can find ">> Engineer to Relational Mode" button among other buttons for logical diagram.
    Philip

  • Working with dates

    hello,
    I am working with dates. It looks like so:
    1st I want to calculate a time span for example calculate cell A2 - cell A1. In cell A1 it is written 01.01.2010 (so first of January) and in cell A2 01.02.2010 (so first of February). I wrote the dates in letters, too, to avoid misunderstandings due to different national date formats.
    Since January has 31 days, the obvious result thereof is "31D". So far so good.
    but now I want to multiply the 31 with a numeric value, but not the 31D, because if I do so, and let's say the numeric value is 3, then my result is 93 DAYS rather than just 93.
    So my question is, is there any possibility how I can further calculate with a result coming from date calculations without having the new result also in the date format? Anything that escapes the "DAYS" in further calculations??
    Thx

    mac-a-rooney wrote:
    I am working with dates.
    1st I want to calculate a time span for example calculate cell A2 - cell A1. In cell A1 it is written 01.01.2010 (so first of January) and in cell A2 01.02.2010 (so first of February). I wrote the dates in letters, too, to avoid misunderstandings due to different national date formats.
    Since January has 31 days, the obvious result thereof is "31D". So far so good.
    but now I want to multiply the 31 with a numeric value, but not the 31D, because if I do so, and let's say the numeric value is 3, then my result is 93 DAYS rather than just 93.
    Are you sure you're asking in the right forum?
    Using Numbers '08, with the following entries:
    B2: Jan 1, 2010
    C2: Feb 1,2010
    D2: =C2-B2
    E2: =3*D2
    I get 31 (not 31D) in D2 and 93 (not 93 DAYS) in D2.
    From your description, I think you're actually using Numbers '09, and Jerry's advice above applies.
    Regards,
    Barry

  • What does HTML have over PDFs when working with data?

    I'm doing some research for a client whose company is moving from PDFs to HTML for their in-house user interfaces.
    What does HTML have over PDFs when working with data?
    Thanks!
    Luke

    PDFs can indeed work with data and can be programmed with javascript. You can do some pretty interesting things with it. It is even possible to create interactive forms on the web using PDF, however it requires server-side support. As a general rule, though, PDFs are terrible as a web interface and it's far easier to work with HTML and PHP.
    HTML is lightweight and PDF isn't (in case that seems like a small thing, it's actually a big negative for PDFs). The success of your PDFs will depend on the versions of acrobat your users use, and getting data in and out of the PDFs will require learning far more about Acrobat's FDF format and XML implementation than you may want to know.

  • Improving the LabVIEW Help: Working with Data Types

    Have you ever had trouble figuring out how to work with the waveform data type, the dynamic data type, or some of the other more complex data types in LabVIEW? As a tech writer on the LabVIEW team, I'd like to improve our documentation about working with data. What would you like to see? What would be helpful? Have you ever given up on a particular data type because it didn't work? Have you ever created a replacement data type because you preferred not to use a LabVIEW data type?
    Lacy Klosterman Rohre | Marketing Editor | National Instruments | 512.683.6376 | ni.com/newsletter

    I've been programming LabVIEW since around 3.1 / 4.0.  Over the years and versions, I've found it necessary to approach a lot of the new datatypes and algorithms with some healthy skepticism because I don't think we're usually given a "fair and balanced" overview.  Some of the things that are highly promoted that bring "ease of use" for beginners (and I *do* understand the importance of that when you're growing your LabVIEW user base) can carry significant performance penalties that are mentioned in more of a whisper, if at all.  I'd just like a more full disclosure of the tradeoffs -- the good, the bad, and the ugly.
    Waveforms -- very rarely use them, unless required for analysis functions.  I have a vague notion (between a belief and a memory) that they used to carry a significant performance penalty compared to arrays, but that the gap is now much smaller.  Still, I'm most comfortable with my old habit of using data arrays.  Some of this came about because a lot of my work uses counters where I may have a variable "dt" value or where my AI is clocked by a counter and the "dt" isn't internally generated.
    Dynamic signals -- have never used them.  Have no clear idea what good they're supposed to be, and they seem to be tied in primarily with Express VI's, which I have also studiously avoided.
    Variants -- Use them only slightly.  Performance issues are a question mark.  Exception: have learned that the implementation of Variant properties allows their use as an efficient way to create associates for string lookup tables. 
    Digital Waveform -- have typically avoided it except when graphing digital data during debug.  Haven't found any compelling reasons to use it.  I do like the notion of a compress/expand capability for sparse digital data, but haven't exercised it enough to trust the implementation. 
    Recap:  I don't need more simple "how-to's" in the help.  I need a LOT more "why bother's" that include both pros and cons. 
    -Kevin P.

  • How to get hebrew characters to work with data merge?

    I'm trying to work with data merge with Hebrew characters and get gibrish on the panel, merging and export.
    I've tried to change the CSV file to Unicode and change the language setting but it still don't work.
    I've worked with data merge before in English and it work perfectly.
    Any Ideas? is this a bug? software constrains?

    Try this:
    Save your file as a UTF-16 BE (Big Endian) file.
    Import showing options in ID. I'm on a PC, choose the below regardless.
    Should look like this in ID.
    Apologies to Farsi-speaking people everywhere. I pulled some text out of a Farsi text file to make up this tab-delimited merge file. I don't speak it, so it is likely servely non-sensical.
    The text editor being used here (first screen shot) is the OpenSource NotePad++. I am also not using the ME version of ID.
    Take care, Mike

  • Will Mountain Lion work with old models of Time Capsule?

    I was using an online backup service but it slowed down my machine so I want to go to an external drive backup. I have upgraded to Mountain Lion and I am thinking of purchasing an older version of Time capsule for my backups. My question is will mountain lion work with older models of Time Capsule? thanks so much.....

    Yes. I have a first generation Time Capsule and it works fine.
    Just make sure you buy at least 3 times the GB you are trying to backup. More is better.

  • Will a charger for a model 1280 work with my model A1278?

    Will a charger for a model 1280 work with my model A1278? Will any L tip 60w work?

    Many thanks for your help.

  • Problem with Data Model and Analysis View

    I create an analyze in BI Publisher and then i create a data model using this object.
    When i try to generate an XML with a number of rows the BI Publisher return an empty XML (Only with de DATA_DS tags but no data).   To bypass this problem i make and XML file by hand and this allow me to create reports and design it but when i try to view the reports i got the message that says "No Data Found".
    So i was check the analysis and all looks appears to be fine,  In the results tab it show me a complete table with the data i was looking to use.
    So i try to repeat the error and when  i try to create the XML for the Data Model i found this two error in the logs:
    [root@server ~]# [2013-07-17T16:37:22.844-04:00] [bi_server1] [WARNING] [] [oracle.xdo] [tid: 2361] [userId: <anonymous>] [ecid: ad7bb40a72b553c0:-3e5f91c5:13ecd037992:-8000-00000000000e2b34,0] [APP: bipublisher#11.1.1] Incomplete xslt._XDONFSEPARATORS: decimal separator: null, grouping separator: null
    [2013-07-17T16:37:26.828-04:00] [bi_server1] [WARNING] [] [oracle.xdo] [tid: 2361] [userId: <anonymous>] [ecid: ad7bb40a72b553c0:-3e5f91c5:13ecd037992:-8000-00000000000e2b3a,0] [APP: bipublisher#11.1.1] Incomplete xslt._XDONFSEPARATORS: decimal separator: null, grouping separator: null
    [2013-07-17T16:37:26.865-04:00] [bi_server1] [WARNING] [] [oracle.xdo] [tid: 2361] [userId: <anonymous>] [ecid: ad7bb40a72b553c0:-3e5f91c5:13ecd037992:-8000-00000000000e2b3a,0] [APP: bipublisher#11.1.1] oracle.xdo.servlet.CreateException: Path: /FOLDER/MODEL.xdm is not pointing to a report. Actual type: ReportItem, sub-type: DataModel[[
            at oracle.xdo.servlet.ReportException.fillInStackTrace(ReportException.java:124)
            at java.lang.Throwable.<init>(Throwable.java:196)
            at java.lang.Exception.<init>(Exception.java:41)
            at oracle.xdo.servlet.ReportException.<init>(ReportException.java:36)
            at oracle.xdo.servlet.CreateException.<init>(CreateException.java:18)
            at oracle.xdo.servlet.ReportRepository.getReport(ReportRepository.java:104)
            at oracle.xdo.servlet.ReportRepository.getReport(ReportRepository.java:128)
            at oracle.xdo.servlet.dataengine.DataProcessorFactory.getDataModelPath(DataProcessorFactory.java:207)
            at oracle.xdo.servlet.dataengine.DataProcessorFactory.isSemanticLayerDataModel(DataProcessorFactory.java:99)
            at oracle.xdo.servlet.dataengine.DataProcessorFactory.isSemanticLayerDataModel(DataProcessorFactory.java:78)
            at oracle.xdo.servlet.ReportModelContextImpl.getReportXMLData(ReportModelContextImpl.java:157)
            at oracle.xdo.servlet.CoreProcessor.process(CoreProcessor.java:346)
            at oracle.xdo.servlet.CoreProcessor.generateDocument(CoreProcessor.java:101)
            at oracle.xdo.servlet.ReportImpl.renderBodyHTTP(ReportImpl.java:1074)
            at oracle.xdo.servlet.ReportImpl.renderReportBodyHTTP(ReportImpl.java:639)
            at oracle.xdo.servlet.XDOServlet.writeReport(XDOServlet.java:492)
            at oracle.xdo.servlet.XDOServlet.writeReport(XDOServlet.java:462)
            at oracle.xdo.servlet.XDOServlet.doGet(XDOServlet.java:280)
            at oracle.xdo.servlet.XDOServlet.doPost(XDOServlet.java:313)
            at javax.servlet.http.HttpServlet.service(HttpServlet.java:727)
            at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
            at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:227)
            at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:125)
            at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:300)
            at weblogic.servlet.internal.TailFilter.doFilter(TailFilter.java:26)
            at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
            at oracle.xdo.servlet.metadata.track.MostRecentFilter.doFilter(MostRecentFilter.java:64)
            at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
            at oracle.xdo.servlet.security.SecurityFilter.doFilter(SecurityFilter.java:125)
            at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
            at oracle.xdo.servlet.init.InitCheckingFilter.doFilter(InitCheckingFilter.java:63)
            at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
            at oracle.security.jps.ee.http.JpsAbsFilter$1.run(JpsAbsFilter.java:119)
            at java.security.AccessController.doPrivileged(Native Method)
            at oracle.security.jps.util.JpsSubject.doAsPrivileged(JpsSubject.java:315)
            at oracle.security.jps.ee.util.JpsPlatformUtil.runJaasMode(JpsPlatformUtil.java:442)
            at oracle.security.jps.ee.http.JpsAbsFilter.runJaasMode(JpsAbsFilter.java:103)
            at oracle.security.jps.ee.http.JpsAbsFilter.doFilter(JpsAbsFilter.java:171)
            at oracle.security.jps.ee.http.JpsFilter.doFilter(JpsFilter.java:71)
            at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
            at oracle.dms.servlet.DMSServletFilter.doFilter(DMSServletFilter.java:139)
            at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
            at oracle.security.jps.ee.http.JpsAbsFilter$1.run(JpsAbsFilter.java:119)
            at java.security.AccessController.doPrivileged(Native Method)
            at oracle.security.jps.util.JpsSubject.doAsPrivileged(JpsSubject.java:315)
            at oracle.security.jps.ee.util.JpsPlatformUtil.runJaasMode(JpsPlatformUtil.java:442)
            at oracle.security.jps.ee.http.JpsAbsFilter.runJaasMode(JpsAbsFilter.java:103)
            at oracle.security.jps.ee.http.JpsAbsFilter.doFilter(JpsAbsFilter.java:171)
            at oracle.security.jps.ee.http.JpsFilter.doFilter(JpsFilter.java:71)
            at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
            at weblogic.servlet.internal.RequestEventsFilter.doFilter(RequestEventsFilter.java:27)
            at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
            at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.wrapRun(WebAppServletContext.java:3715)
            at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3681)
            at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
            at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:120)
            at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:2277)
            at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:2183)
            at weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1454)
            at weblogic.work.ExecuteThread.execute(ExecuteThread.java:209)
            at weblogic.work.ExecuteThread.run(ExecuteThread.java:178)
    And when i try to view the report that use the analysis i got this two warning in the logs:
    [2013-07-17T16:58:01.615-04:00] [bi_server1] [WARNING] [] [oracle.xdo] [tid: 57] [userId: <anonymous>] [ecid: ad7bb40a72b553c0:-3e5f91c5:13ecd037992:-8000-00000000000e2d7c,0] [APP: bipublisher#11.1.1] Incomplete xslt._XDONFSEPARATORS: decimal separator: null, grouping separator: null
    [2013-07-17T16:58:02.034-04:00] [bi_server1] [WARNING] [] [oracle.xdo] [tid: 57] [userId: <anonymous>] [ecid: ad7bb40a72b553c0:-3e5f91c5:13ecd037992:-8000-00000000000e2d84,0] [APP: bipublisher#11.1.1] Incomplete xslt._XDONFSEPARATORS: decimal separator: null, grouping separator: null
    As i understand there is has a reference to a null value but i cant find what column of the analysis has the problem.
    Any ideas about how to solve or debug this?
    Thanks

    I follow your instructions and it works fine. I can create the XML and the reports can show data.
    So, i already know that there is no problem with the data and with the BI Publisher installation but i still doesn know what is the problem with the analysis view that fails.
    Any idea how to debug it?
    Thanks.

  • Need help with Data Model for Private Messaging

    Sad to say, but it looks like I just really screwed up the design of my Private Messaging (PM) module...  *sigh*
    What looked good on paper doesn't seem to be practical in application.
    I am hoping some of you Oracle gurus can help me come up with a better design!!
    Here is my current design...
    member -||-----0<- private_msg_recipient ->0------||- private_msg
    MEMBER table
    - id
    - email
    - username
    - first_name
    PRIVATE_MSG_RECIPIENT table
    - id
    - member_id_to
    - message_id
    - flag
    - created_on
    - updated_on
    - read_on
    - deleted_on
    - purged_on
    PRIVATE_MSG table
    - id
    - member_id_from
    - subject
    - body
    - flag
    - sent_on
    - updated_on
    - sender_deleted_on
    - sender_purged_on
    ***Short explanation of how the application currently works...
    - Sender creates a PM and sends it to a Recipient.
    - The PM appears in the Sender's "Sent" folder in my website
    - The PM also appears in the Recipient's "Incoming" folder.
    - If the Recipient deletes the PM, I set "deleted_on" and my code moves the PM from Recipient's "Inbox" to the "Trash" folder.  (Record doesn't actually move!)
    - If the Recipient "permanently deletes" the PM from his/her "Trash", I set "purged_on" and my code removes the PM from the Recipient's Message Center.  (Record still in database!)
    - If the Sender deletes the PM, I set "sender_deleted_on" and my code moves the PM from the Sender's "Sent" folder to the "Trash" folder.  (Record doesn't actually move!)
    - If the Recipient "permanently deletes" the PM from his/her "Trash", I set "sender_purged_on" and my code removes the PM from the Sender's Message Center.  (Record still in database!)
    Here are my problems...
    1.) I can't store PM's forever.
    2.) Because of my design, the Sender really owns the PM, and if I add code to REMOVE the PM from the database once it has a "sender_purged_on" value, then that would in essence remove the PM from the Recipient's Inbox as well!!
    In order to remove a PM from the database, I would have to make sure that *both* the Recipient has "purged_on" value and the Sender has a "sender_purged_on" value.  (Lot's of Application Logic for something which should be simple?!)
    I am wondering if I need to change my Data Model to something that allows my autonomy when it comes to the Sender and/or the Recipient deleting the PM for good...
    One the other hand, I believe I did a good job or normalizing the data.  And my current Data Model is the most efficient when it comes to saving storage space and not having dups.
    Maybe I do indeed just need need to write application logic - or a cron job - which checks to make sure that *both* the Sender an Recipient have deleted the PM before it actually flushes it out of my database to free up space?!
    Of course, if one party sits on their PM's forever, then I can never clear things out of my database to free up space...
    What should I do??
    Some expert advice would be welcome!!
    Sincerely,
    Debbie

    rp0428,
    I think I am starting to see my evil ways and where I went wrong... 
    > Unfortunately his design is just as denormalized as yours
    I see that now.  My bad!!
    > the last two columns have NOTHING to do with the message itself so do NOT belong in a normalized table.
    > And his design:
    >
    > Same comment - those last two columns also have NOTHING to do with the message itself.
    Right.
    > The message table should just have columns directly related to the message. It is a list of unique messages: no more, no less.
    Right.
    > Mark gave you hints to the proper normalized design using an INTERSECT table.
    > that table might list: sender, recipient, sender_delete_flag, recipient_delete_flag.
    > As mark suggested you could also have one or two DATEs related to when the delete flags were set. I would just make the columns DATE fields.
    >
    > Once both date columns have a value you can delete the message (or delete all messages older than 30+ days).
    >
    > When both flags are set you can delete the message itself that references the sender and the message sent.
    Okay, how does this revised design look...
    MEMBER --||-----0<-- PM_DISTRIBUTION -->0-------||-- PRIVATE_MSG
    MEMBER table
    - id
    - email
    - username
    - first_name
    and so on...
    PM_DISTRIBUTION table (Maybe you can think of a better name??)
    - id
    - private_msg_id
    - sender_id
    - recipient_id
    - sender_flag
    - sender_deleted_on
    - sender_purged_on
    - recipient_flag
    - recipient_read_on
    - recipient_deleted_on
    - recipient_purged_on
    PRIVATE_MSG
    - id
    - subject
    - body
    - sent_on
    Is that what you were describing to me?
    Quickly reflecting on this new design...
    1.) It should now be in 3rd Normal Form, right?
    2.) It should allow the Sender and Recipient to freely and independently "delete" or "purge" a PM with no impact on the other party, right?
    Here are a few Potential Issues that I see, though...
    a.) What is to stop there from being TWO SENDERS of a PM?
    In retrospect, that is why I originally stuck "member_id_from" in the PRIVATE_MSG table!!  The logic being, that a PM only ever has *one* Sender.
    I guess I would have to add either Application Logic, or Database Logic, or both to ensure that a given PM never has more than one Sender, right?
    b.) If the design above is what you were hinting at, and if it is thus "correct", then is there any conflict with my Business Rule: "Any given User shall only be allowed 100 Messages between his/her Incoming, Sent and Trash folders."
    Because the Sender is no longer "tightly bound" to the PRIVATE_MSG, in my scenario above...
    Debbie could send 100 PM's, hit her quota, then turn around and delete and purge all 100 Sent PM's and that should in no way impact the 100 PM's sitting in other Users' Inboxes, right??
    I think this works like I want...
    Sincerely,
    Debbie

  • Help Needed with Data-modeling to build an application on

    Hi would anyone be able to help me in creating a data model  cause im really stuck with this one .Basically if been asked to create a survey application in oracle apex that use to excel based . So the info i was given was in a form of  excel sheet which looks like this
    NAME
    E-MAIL
    TSSA
    ORACLE
    HP
    IBM
    MS
    SAP
    INTERGRAPH
    CISCO
    Relationship
    Contracting
    Performance
    Architecture
    Supplier Feedback
    comments
    Jxxxxxx yyyyyyf
    [email protected]
    Yes
    Yes
    Yes
    Yes
    x
    requested to be added
    nnnitha iiiiiah
    [email protected]
    Yes
    Yes
    Yes
    x
    x
    Knnnn kkkikot
    [email protected]
    Yes
    x
    x
    is not payed
    Gggrt Louuuue
    [email protected]
    Yes
    Yes
    Yes
    Yes
    Yes
    Yes
    Yes
    Yes
    x
    x
    x
    x
    jeiiiha ad
    [email protected]
    Yes
    x
    to meet with
    John Rat
    [email protected]
    Yes
    x
    x
    So where it says yes thous are the vendors that people associated with them have to asses and where there's an X thous are the topics that the vendors have to be rated on . So if for example the first guy on the list Jxxxxxx yyyyyyf will asses TSSA , ORACLE, HP , IBM , MS , SAP  on the topic of Architecture and if you look at the second user nnnitha iiiiiah he would rate TSSA , ORACLE , INTERGRAPH on the topics of Relationship and performance  . Any idea how i could data model this to get my table structures right .so that features like completion status could be displayed to the user through APEX which can only be done by a correct data-model i have tried normalization but  i did go anywhere becauce there are so many variations any idea on how you would go about data modeling this would be greatly appreciated thank you    

    Not really an APEX specific question..  Maybe you should try posting this in the data modeler forum : SQL Developer Data Modeler
    Thank you,
    Tony Miller
    LuvMuffin Software

  • GPS does not appear to work with data-roaming off and no wi-fi network

    I have been using an application called imarkmyspot to produce way-points for synchronising with my camera. When I am in my home country this works fine and a GPS location is found quickly.
    I've recently been abroad and had data-roaming turned off to avoid high expense. The point is that in this situation the GPS does not work (with imarkmyspot or Goggle Maps) unless I can access an open wi-fi network.
    So my question is, is this correct? Does GPS only work with either data access on or when within range of an open wi-fi network.
    If this is feature (bug) of the iPhone then this makes applications such as imarkmyspot rather useless when travelling. Alternatively it could be imarkmyspot that has this problem of course. But I'm quite sure Google Maps did not work either in this situation either.

    The iPhone's GPS does work without a data connection.
    Two weeks ago on a work-related trip I knew I'd be traveling a rural highway that would leave me without data/voice from any source for at least 45 minutes (no GSM service). Before leaving a city with data, I zoomed out using Google Maps far enough to show most of my route, then "nudged" the map so the application wouldn't try to get a data update for the map. I simply used the top power button to turn on and off the screen. The iPhone and GMaps tracked me through for the entire blacked-out section of my trip - I made sure the device had a decent view of the sky and it got a fix and nailed the location within a few seconds.
    Ironically, I'm returning that device in the next few hours for an exchange. The GPS hardware, it seems, has failed about a week ago - the device gets A-GPS fixes now (it shows the nearby 3G and EDGE towers), but won't lock in for more than a second or two. My iPhone's GPS worked, and now it doesn't - AppleCare told me my GPS hardware has failed - and you're may be exhibiting the same failure.
    Some A-GPS devices use their sources to get close by - like my TomTom GO and the ephemeris files it grabs from their servers to get updates on where satellite locations are - and then use the on-board chip to nail the device's position. From what I understand of the iPhone's assistance mechanism, the data tower's and Wi-Fi physical location needs to be correct and available in an accessible database somewhere - my iPhone was showing me near Houston TX (2200 mi/3500 km distant from my actual location) when using 3G data until the data source was updated.

  • Can't create connection with Data Modeler 2.0.0 570

    I have installed SQL Developer Data Modeler on my Windows Vista Home Premium system and am unable to create a connection to import a Data Dictionary. When I select TNS from the connection type dropdown list , I get a message below that states: 'Warning ORACLE_HOME environment variable is not set'.
    This is odd, since I use SQL Developer daily on this same system with no issues. I have checked with regedit and the ORACLE_HOME environment variable is indeed set.
    Any ideas?

    I ran into a similar where Data Modeler was finding the TNSNAMES.ORA file for the older Oracle 8i client (which I have to keep installed because some older legacy software depends on it) rather than the 10g client. My solution was to make use of a suggestion for a simialr problem with SQL Developer and create a DOS "cmd" file in the "+<data modeler install dir>+\datamodeler" directory that sets the ORACLE_HOME environment variable to my 10g client directory and then runs "datamodeler.exe". My file is named "datamodeler.cmd" and contains the following two lines:
    SET ORACLE_HOME=C:\oracle\product\10.2.0\client_1
    start datamodeler.exe
    HTH.
    Ed. H.

  • Import of DDL with data modeler

    Hello,
    I begin to use the data modeler of sql developer. I create un relational model with constraints about range and values list. Then I generate the corresponding DDL :
    -- Generated by Oracle SQL Developer Data Modeler Version: 2.0.0 Build: 584
    -- at: 2010-03-08 19:26:07
    -- site: Oracle Database 10g
    -- type: Oracle Database 10g
    CREATE TABLE ACTLIST
    TIME_ACT INTEGER NOT NULL ,
    NON_CTL_AREA_NAME VARCHAR2 (12) NOT NULL ,
    AIRSPACE_ENV_NAME VARCHAR2 (20) NOT NULL
    ) LOGGING
    ALTER TABLE ACTLIST
    ADD CONSTRAINT TIME_ACT_CHCK
    CHECK ( TIME_ACT BETWEEN 1 AND 2000)
    CREATE TABLE AIRCR_MO
    AIRCRAFT_NAME VARCHAR2 (4) NOT NULL ,
    WAKE_TURBULENCE_CAT VARCHAR2 (1) NOT NULL ,
    TAS INTEGER NOT NULL
    ) LOGGING
    ALTER TABLE AIRCR_MO
    ADD CONSTRAINT WAKE_TURBULENCE_CAT_CHCK
    CHECK ( DESCRIPTION IN ( 'L', 'M', 'H' ))
    ALTER TABLE AIRCR_MO
    ADD CONSTRAINT TAS_CHCK
    CHECK ( TAS BETWEEN 0 AND 999)
    Now, I import this DDL and constraints about range and values list are lost. Is it normal ? How Can I do to preserve these data ? Domains on each columns are effectively created but raneg and values list are empty !
    Thank you

    Hello,
    these definitions are generated as check constraint:
    ALTER TABLE AIRCR_MO
    ADD CONSTRAINT WAKE_TURBULENCE_CAT_CHCK
    CHECK ( DESCRIPTION IN ( 'L', 'M', 'H' )) and they are imported as check constraint in current implementation. We have ER logged for that - to import them as list of ranges or/and list of values.
    Philip

Maybe you are looking for