Offline data store related

Hi,
i Have created application and its deployed. now operator has done the entry into the form and data stored in tables.
where the datafile will stored in drive.what will be the path. where i can see table data stored file in drive like c or d.
please reply with an example.hope i have explained well.
Thanks & Regards
Nisha.

Hi,
I think this is not relating to Apex.
You need check from DB parameters where data files are stored
And it might good if you read database concept
http://download.oracle.com/docs/cd/B19306_01/server.102/b14220/physical.htm
Regards,
Jari

Similar Messages

  • What do you recommend to use as an offline data store, since SQL CE support is not in VS 2013?

    A few years back I was architecting an occasionally connected .Net desktop application. 
    VS 2010 was offering full support for Microsoft Sync Framework and SQL CE with Entity Framework. 
    This seemed like the perfect marriage, so I ran with it, and the resulting software solution is still successfully running in production, years later. 
    Jump forward to today, and I am architecting a new occasionally connected .Net desktop application. 
    I was really looking forward to taking advantage of the advances made by Microsoft in using the tools built into VS 2013. 
    However, what I discovered has dumbfounded me.  VS 2013 has no designer support for Sync Framework, and worse, built in support for SQL CE has been completely removed, including the ability to generate Entity Framework models from a
    CE database using the designer. 
    My question to the community is, what tools should I be using to solve the problem of offline storage in my brand new .Net application? 
    I am aware of ErikEJ’s SQL Server Compact Toolbox, which brings back some support for these features in VS 2013, but it is not as fully featured as the VS 2010 native support was, plus it does not have the institutional “Microsoft” stamp on it. 
    I am building a multimillion dollar corporate solution that I will have to support for many years.
     I would like to have some comfort that the technologies I select, today, will still be supported 5 years from now, unlike the way Microsoft has discontinued supporting Sync Framework and CE in the most recent VS. 
    I can accept open source technologies, because there is a community behind them, or off the shelf corporate solutions, since they will be driven by financial gain, but I have trouble committing to a solution that is solely supported by an individual,
    even if that person is a very talented Microsoft MVP.
    Some of the features of SQL CE that I would like to keep are
    Built in encryption of the file on disk
    Easy querying with an ORM, like Entity Framework
    Tools to easily sync up the offline data store with values from SQL Server (even better if this can be done remotely, such as over WCF)
    Does not require installation of additional software on the client machine, as SQL Express would
    Please, provide your feedback to let me know how you have achieved this, without resorting to simply using an older version of VS or Management Studio. 
    Thank you.

    Hello,
    Based on your description, you can try to use SQL Server 2012 Express LocalDB.
    LocalDB is created specifically for developers. It is very easy to install and requires no management, but it offers the same T-SQL language, programming surface and client-side providers as the regular SQL Server Express.
    SQL Server LocalDB can work with Entity Framework and ADO.NET Syc Framework. However, there is no built-in encryption feature in LocalDB which
    can let you encrypt database. You should decrypt/encrypt data on your own, for example, using
    Cryptographic Functions
    Reference:SQL Express v LocalDB v SQL Compact Edition
    Regards,
    Fanny Liu
    If you have any feedback on our support, please click here.
    Fanny Liu
    TechNet Community Support

  • What is the event by which i can get that user deny the offline data store when the message comes in the browser

    I have an application which stores manifest for offline use. Now firefox asking for a confirm message that if i want to allow or not. If i say 'allow' the application works fine. but if i say 'not now' or 'never for this site' the application stops. What is the event to check that user deny to allowing store data. i used different option like - cached,noupdate,error,uncached,idle,obsolete. as an example- applicationCache.addEventListener('idle', function () {});
    but none of this says that user deny to allow. Please tell me the option that i can check that user does not want to allow and do my rest work.

    It should be possible to allow offline data storage for only certain websites, and that setting is in Options > Advanced > Network. If the answer is "don't do it", then it should not be done. Is that not what is happening?

  • Error while viewing data from data store

    Hello Gurus,
    We are facing issue with driver when we try to view data from a data store related to Hyperion Essbase technology.
    ODI version is 11.1.1.6.
    Following is the error that we are getting:
    java.lang.IllegalArgumentException: Driver name cannot be empty
         at org.springframework.util.Assert.hasText(Assert.java:161)
         at com.sunopsis.sql.SnpsConnection.setDriverName(SnpsConnection.java:302)
         at com.sunopsis.dwg.dbobj.DwgConnectConnection.setDefaultConnectDefinition(DwgConnectConnection.java:380)
         at com.sunopsis.dwg.dbobj.DwgConnectConnection.<init>(DwgConnectConnection.java:274)
         at com.sunopsis.dwg.dbobj.DwgConnectConnection.<init>(DwgConnectConnection.java:288)
         at oracle.odi.core.datasource.dwgobject.support.DwgConnectConnectionCreatorImpl.createDwgConnectConnection(DwgConnectConnectionCreatorImpl.java:53)
         at com.sunopsis.graphical.frame.edit.EditFrameTableData.snpsInitializeSnpsComponentsSpecificRules(EditFrameTableData.java:85)
         at com.sunopsis.graphical.frame.SnpsEditFrame.snpsInitialize(SnpsEditFrame.java:1413)
         at com.sunopsis.graphical.frame.edit.AbstractEditFrameGridBorland.initialize(AbstractEditFrameGridBorland.java:623)
         at com.sunopsis.graphical.frame.edit.AbstractEditFrameGridBorland.<init>(AbstractEditFrameGridBorland.java:868)
         at com.sunopsis.graphical.frame.edit.EditFrameTableData.<init>(EditFrameTableData.java:50)
         at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
         at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
         at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
         at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
         at oracle.odi.ui.editor.AbstractOdiEditor$1.run(AbstractOdiEditor.java:176)
         at oracle.ide.dialogs.ProgressBar.run(ProgressBar.java:655)
         at java.lang.Thread.run(Thread.java:662)Is there any specific JAR file related to Hyperion Essbase ?
    and where do we find the default drivers that come with ODI?
    Please help.
    Thanks,
    Santy.

    You cannot view the data from an essbase data store as it isn't configured with a jdbc driver that supports this function

  • Offline Data Transaction Semantics and Attachments

    I'm researching migrating an existing web and mobile application to Azure. Currently, we're using a NoSQL solution that syncs to SQLite on the client and maps local documents to model subclasses similar to Core Data in iOS. Offline support is a key differentiation
    in our market, so I'm focused on the detailed mechanics of sync. 
    One of our current pain points is that documents synced on both on the server and client do not arrive in a transaction. To use an example, if the iOS client created 10 Projects, each with 5 related Tasks, the first Project might be synced to the server
    and trigger an event to process its related Tasks before the remaining Projects were inserted and the Tasks table even started syncing. This would cause processing to fail as the related Tasks were not yet available in the database. Since it appears that syncing
    rows, some of which had conflicts, can result in only some of the rows being inserted, I'm under the impression server side inserts do not occur in a transaction, even at the scope of a table. Would this be accurate conclusion? If not, what are the transaction
    semantics with offline sync?
    Also, is there a strategy for syncing related binary content offline? For example, in our current system, each document can have n-number of media attachments that are synced as part of a multi-part request. However, it appears we would either need to store
    related media as blobs in a table column or roll our own method of synchronizing related files as rows were pushed and pulled if created on the device while offline.  
    Thanks, 
    - Scott

    Hi
    Thanks for the update.
    Good to know the different ways to do it.
    Not sure if this fits in workflow forum too. as the question is primarily  around the  data handling and storage of attachments and documents.
    Hence I close this thread in this forum, And created the thread in 'Workflow forum ' to get to know the response.
    New Thread : Webforms - data handling and storage
    Please update your thoughts in the new thread. Appreciate your time and effort.
    Thanks alot.
    Best Regards
    Saujanya
    Edited by: Saujanya on Nov 17, 2011 12:20 PM

  • How to select from multiple tables which reside on different data stores ?

    Suppose I have two data stores in one TimesTen instance:
    1) Datastore A:
    table1
    2) Datastore B:
    table2
    I want to make a query like this:
    select ... from table1, table2 where table1.colA = table2.colB
    Can I ? If not, is there a workaround ?
    BTW, because of business, we have to use two or more different datastores, so we can not put table1/table2 in the same datastore.
    Thanks very much.

    You can query multiple TimesTen databases, but your original question was about joining tables from two databases, which is not supported.
    Using Cache Connect to Oracle to query an Oracle database is not distributed. It's still one single Oracle database you are querying. You cannot join a table in the TimesTen database with a table in the Oracle database, this is not allowed.
    If you are willing to share your business requiremens, we can take a look and see what solution might work for you. Would you like to discuss this offline?
    Susan

  • 703: Subdaemon connect to data store failed with error TT9999

    All,
    I'm getting the following error whilst trying to connect to a TimesTen DB:
    connect "DSN=my_cachedb";
    703: Subdaemon connect to data store failed with error TT9999
    In the tterrors.log:
    16:39:24.71 Warn: : 2568: 3596 ------------------: subdaemon process exited
    16:39:24.71 Warn: : 2568: 3596 exited while connected to data store '/u01/ttdata/datastores/my_cachedb' shm 33554529 count=1
    16:39:24.71 Warn: : 2568: daRecovery: subdaemon 3596, managing data store, failed: invalidate (failcode=202)
    16:39:24.71 Warn: : 2568: Invalidating the data store (failcode 202, recovery for 3596)
    16:39:24.72 Err : : 2568: TT14000: TimesTen daemon internal error: Could not send 'manage' request to subdaemon rc -2 err1 703 err2 9999
    16:39:24.72 Warn: : 2568: 3619 Subdaemon reports creation failure
    16:39:24.72 Err : : 2568: TT14000: TimesTen daemon internal error: Deleting 3619/0x1558650/'/u01/ttdata/datastores/my_cachedb' - from association table - not found
    16:39:24.72 Err : : 2568: TT14004: TimesTen daemon creation failed: Could not del from dbByPid internal table
    16:39:24.81 Warn: : 2568: child process 3596 terminated with signal 11
    16:39:25.09 Err : : 2568: TT14000: TimesTen daemon internal error: daRecovery for 3619: No such data store '/u01/ttdata/datastores/my_cachedb'
    I've checked and the datastore does exist and is owned by the timesten UNIX user.
    ttversion:
    TimesTen Release 11.2.2.2.0 (64 bit Linux/x86_64) (tt1122:53396) 2011-12-23T09:26:28Z
    Instance admin: timesten
    Instance home directory: /home/timesten/TimesTen/tt1122
    Group owner: timesten
    Daemon home directory: /home/timesten/TimesTen/tt1122/info
    PL/SQL enabled.
    Datastore definition from sys.odbc.ini:
    [my_cachedb]
    Driver=/home/timesten/TimesTen/tt1122/lib/libtten.so
    DataStore=/u01/ttdata/datastores/my_cachedb
    LogDir=/u01/ttdata/logs
    PermSize=40
    TempSize=32
    DatabaseCharacterSet=AL32UTF8
    OracleNetServiceName=testdb
    Kernel parameters from sysctl -a:
    kernel.shmmax = 68719476736
    kernel.shmall = 4294967296
    Memory / SWAP:
    MemTotal: 2050784 kB
    SWAP: /dev/mapper/VolGroup00-LogVol01 partition 4095992
    I'm new to TimesTen and I'm planning on evaluationg it to see if it could solve an issue we're having. Any suggestions would be much appreciated.
    Thanks,
    Ian.

    Hi Ian,
    Can you please answer the following / provide the following information:
    1. What are your kernel parameters relating to semaphores set to? Is anything else on the mahcine using significant numbers of semaphores?
    2. Please provide the output of the following shell commands:
    ls -ld /u01
    ls -ld /u01/ttdata
    ls -ld /u01/ttdata/datastores
    ls -ld /u01/ttdata/logs
    3. Please provide an excerpt of the detailed message log (ttmesg.log) between around 16:38 and 16:40 (i.e. from a little while before the problem until after the problem).
    Thanks,
    Chris

  • Using offline data sync in Mobile Services

    Hi All,
    I am working on offline data sync in Mobile Services . Followed the below URL
    http://azure.microsoft.com/en-gb/documentation/articles/mobile-services-windows-store-dotnet-get-started-offline-data/#enable-offline-app
    Have an issue at store.DefineTable<TodoItem>();
    when i run the method i am getting any exception as 
    System.ArgumentException was unhandled by user code
      HResult=-2147024809
      Message=An item with the same key has already been added.
      Source=mscorlib
      StackTrace:
           at System.ThrowHelper.ThrowArgumentException(ExceptionResource resource)
           at System.Collections.Generic.Dictionary`2.Insert(TKey key, TValue value, Boolean add)
           at System.Linq.Enumerable.ToDictionary[TSource,TKey,TElement](IEnumerable`1 source, Func`2 keySelector, Func`2 elementSelector, IEqualityComparer`1 comparer)
           at System.Linq.Enumerable.ToDictionary[TSource,TKey](IEnumerable`1 source, Func`2 keySelector, IEqualityComparer`1 comparer)
           at Microsoft.WindowsAzure.MobileServices.SQLiteStore.MobileServiceSQLiteStore.DefineTable(String tableName, JObject item)
           at Microsoft.WindowsAzure.MobileServices.SQLiteStore.MobileServiceSQLiteStoreExtensions.DefineTable[T](MobileServiceSQLiteStore store, MobileServiceJsonSerializerSettings settings)
           at Microsoft.WindowsAzure.MobileServices.SQLiteStore.MobileServiceSQLiteStoreExtensions.DefineTable[T](MobileServiceSQLiteStore store)
      private async Task InitLocalStoreAsync()
        if (!App.MobileService.SyncContext.IsInitialized)
            var store = new MobileServiceSQLiteStore("localstore.db");
            store.DefineTable<TodoItem>();
            await App.MobileService.SyncContext.InitializeAsync(store);
        await SyncAsync();
    Need help.
    Thanks,
    Sai.

    Hi,
    The Azure Mobile Services forum is here:
    https://social.msdn.microsoft.com/Forums/en-US/home?forum=azuremobile&filter=alltypes&sort=lastpostdesc
    Don't retire TechNet! -
    (Don't give up yet - 13,225+ strong and growing)

  • Problem with migrating data (SQL Server 2k offline data capture)

    Hi there,
    I try to migrate a SQL Server 2000 database to Oracle 10g (R2) using offline data capture. I following the online Tutorial “Migrate from Microsoft SQL Server to Oracle Database 10g Using Oracle Migration Workbench” and everything seems OK until “Migrating the tablespaces, users and user tables to the destination database” http://www.oracle.com/technology/obe/10gr2_db_vmware/develop/omwb/omwb.htm#t5. I have two errors on 2 (out of 85 ) tables saying “ Failed to create default for Table :
    ORA-00907: missing right parenthesis”. I checked the column/table names and they seems OK. I don’t understand why I got these errors. However, when I checked the ‘SA’ Schema in Oracle Enterprise Manager Console, I can see all the tables (including the two problem ones).
    I then carry on to the next step and tried to migrate data to the destination database (http://www.oracle.com/technology/obe/10gr2_db_vmware/develop/omwb/omwb.htm#t6) I am now stuck on step 5. “…copy the files from c:\omwb\data_files to the c:\omwb\Omwb\sqlloader_scripts\SQLServer2K\<timestamp>\Oracle directory…” I cannot find the ‘c:\omwb\data_files’ directory, in fact, I don’t have a directory called ‘data_files’ on my machine. I noticed the ‘c:\omwb\Omwb\sqlloader_scripts\SQLServer2K\<timestamp>\Oracle’ directory contains the all the 85 [tableName].ctl files (plus two other files: ‘sql_load_script’ and ‘sql_load_script.sh’). From the screenshots online, ‘c:\omwb\data_files’ seems contains files with the same name as the [tableName]. Therefore, I did a search with [tableName] but cannot find any apart from the [tableName].ctl file in the ‘c:\omwb\Omwb\sqlloader_scripts\SQLServer2K\<timestamp>\Oracle’ directory. What should I do next?
    OS: windows 2003 with SP1
    Oracle: 10g Release 2
    SQL Server 2000
    Any help would be extremely appreciated.
    Helen

    Helen,
    Sorry, I am new here. Could you please tell me (or point me to the related documents about) how to output the Oracle model as a script?Action-> Generate Migration Scripts
    And what do you mean by ‘The default conversion may have failed’? Do you mean data type mapping? I went through all the columns and tables and checked the data types in the Oracle model already. The processing for the default for a table column could have something the basic workbench default parser cannot handle.
    I hope you are finding the workbench to be a productive aid to migration.
    Regards,
    Turloch
    Oracle Migration Workbench Team

  • Offline Data Caching Native iOS

    Hi, guy
    I am working on Offline Data SMP 3.0, i am saved Service Document and MetaData in cache already but, when turn off internet connection, I will read service document and metadata in cache but i can not get service document schema and metadata schema because entityType in service document and metadata are null, Please suggest me how to read service document and metadata in Cache on Offline Mode or give me some example.
    Thank you

    You have to leverage cache library to store the service document and metadata document. My expert area is not IOS, I have a sample code for Android. Given below is an example :
    IODataServiceDocument serviceDocument = null;
    try {
    //Getting Service Document from the cache
    serviceDocument =
    (IODataServiceDocument)mApplication.getCache().readDocument(
    DocumentType.ServiceDocument,
    FlightModelConstants.FLIGHT_MODEL_SERVICE_DOCUMENT);
    } catch (CacheException e) {
    Log.v(TAG, e.getLocalizedMessage());
    //Checking if Service Document was not found in cache,
    if (serviceDocument == null) {
    mProgress =
    ProgressDialog.show(this,
    (CharSequence) "",
    (CharSequence)
    getString(R.string.serviceDocumentProgress),
    true);
    mRequestTag = REQUEST_SERVICE_DOCUMENT;
    //Preparing request to retrieve Service Document from the server
    IRequest request =
    RequestBuilder.getInstance().buildServiceDocumentRequest(this,
    REQUEST_SERVICE_DOCUMENT);
    //Sending the request
    mApplication.getRequestManager().makeRequest(request);
    } else {
    Toast.makeText(this, R.string.msgServiceDocumentCache,
    Toast.LENGTH_LONG).show();
    //Service Document was in the cache, now we are going to retrieve the
    Schema
    this.getServiceSchema();
    The APIs are given in bold letters. Hope that helps.
    Code to be added in onSuccess callback method to store the document:
    IODataServiceDocument serviceDocument = parser
    .parseODataServiceDocument(responseString);
    mApplication.getCache().storeDocument(serviceDocument,
    DocumentType.ServiceDocument,
    FlightModelConstants.FLIGHT_MODEL_SERVICE_DOCUMENT);

  • Horizontal Fragmentation in NoSql Data Stores

    Hello every body
    My question is :
    can we apply the optimization technique of " Horizontal fragmentation (partitioning)" in this new type of data stores (NoSql) just like in relational databases?
    thanks at advance

    No, not really. I suggest reading the documentation here:
    http://www.oracle.com/technetwork/products/nosqldb/overview/index.html
    You may want to start with the link "Intelligent Driver" and also read the Getting Started Guide.
    --mark                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

  • Cannot create data store shared-memory segment error

    Hi,
    Here is some background information:
    [ttadmin@timesten-la-p1 ~]$ ttversion
    TimesTen Release 11.2.1.3.0 (64 bit Linux/x86_64) (cmttp1:53388) 2009-08-21T05:34:23Z
    Instance admin: ttadmin
    Instance home directory: /u01/app/ttadmin/TimesTen/cmttp1
    Group owner: ttadmin
    Daemon home directory: /u01/app/ttadmin/TimesTen/cmttp1/info
    PL/SQL enabled.
    [ttadmin@timesten-la-p1 ~]$ uname -a
    Linux timesten-la-p1 2.6.18-164.6.1.el5 #1 SMP Tue Oct 27 11:28:30 EDT 2009 x86_64 x86_64 x86_64 GNU/Linux
    [root@timesten-la-p1 ~]# cat /proc/sys/kernel/shmmax
    68719476736
    [ttadmin@timesten-la-p1 ~]$ cat /proc/meminfo
    MemTotal: 148426936 kB
    MemFree: 116542072 kB
    Buffers: 465800 kB
    Cached: 30228196 kB
    SwapCached: 0 kB
    Active: 5739276 kB
    Inactive: 25119448 kB
    HighTotal: 0 kB
    HighFree: 0 kB
    LowTotal: 148426936 kB
    LowFree: 116542072 kB
    SwapTotal: 16777208 kB
    SwapFree: 16777208 kB
    Dirty: 60 kB
    Writeback: 0 kB
    AnonPages: 164740 kB
    Mapped: 39188 kB
    Slab: 970548 kB
    PageTables: 10428 kB
    NFS_Unstable: 0 kB
    Bounce: 0 kB
    CommitLimit: 90990676 kB
    Committed_AS: 615028 kB
    VmallocTotal: 34359738367 kB
    VmallocUsed: 274804 kB
    VmallocChunk: 34359462519 kB
    HugePages_Total: 0
    HugePages_Free: 0
    HugePages_Rsvd: 0
    Hugepagesize: 2048 kB
    extract from sys.odbc.ini
    [cachealone2]
    Driver=/u01/app/ttadmin/TimesTen/cmttp1/lib/libtten.so
    DataStore=/u02/timesten/datastore/cachealone2/cachealone2
    PermSize=14336
    OracleNetServiceName=ttdev
    DatabaseCharacterset=WE8ISO8859P1
    ConnectionCharacterSet=WE8ISO8859P1
    [ttadmin@timesten-la-p1 ~]$ grep SwapTotal /proc/meminfo
    SwapTotal: 16777208 kB
    Though we have around 140GB memory available and 65GB on the shmmax, we are unable to increase the PermSize to any thing more than 14GB. When I changed it to PermSize=15359, I am getting following error.
    [ttadmin@timesten-la-p1 ~]$ ttIsql "DSN=cachealone2"
    Copyright (c) 1996-2009, Oracle. All rights reserved.
    Type ? or "help" for help, type "exit" to quit ttIsql.
    connect "DSN=cachealone2";
    836: Cannot create data store shared-memory segment, error 28
    703: Subdaemon connect to data store failed with error TT836
    The command failed.
    Done.
    I am not sure why this is not working, considering we have got 144GB RAM and 64GB shmmax allocated! Any help is much appreciated.
    Regards,
    Raj

    Those parameters look ok for a 100GB shared memory segment. Also check the following:
    ulimit - a mechanism to restrict the amount of system resources a process can consume. Your instance administrator user, the user who installed Oracle TimesTen needs to be allocated enough lockable memory resource to load and lock your Oracle TimesTen shared memory segment.
    This is configured with the memlock entry in the OS file /etc/security/limits.conf for the instance administrator.
    To view the current setting run the OS command
    $ ulimit -l
    and to set it to a value dynamically use
    $ ulimit -l <value>.
    Once changed you need to restart the TimesTen master daemon for the change to be picked up.
    $ ttDaemonAdmin -restart
    Beware sometimes ulimit is set in the instance administrators "~/.bashrc" or "~/.bash_profile" file which can override what's set in /etc/security/limits.conf
    If this is ok then it might be related to Hugepages. If TT is configured to use Hugepages then you need enough Hugepages to accommodate the 100GB shared memory segment. TT is configured for Hugepages if the following entry is in the /u01/app/oracle/EXALYTICS_MWHOME/TimesTen/tt1122/info/ttendaemon.options file:
    -linuxLargePageAlignment 2
    So if configured for Hugepages please see this example of how to set an appropriate Hugepages setting:
    Total the amount of memory required to accommodate your TimesTen database from /u01/app/oracle/EXALYTICS_MWHOME/TimesTen/tt1122/info/sys.odbc.ini
    PermSize+TempSize+LogBufMB+64MB Overhead
    For example consider a TimesTen database of size:
    PermSize=250000 (unit is MB)
    TempSize=100000
    LogBufMB=1024
    Total Memory = 250000+100000+1024+64 = 351088MB
    The Hugepages pagesize on the Exalytics machine is 2048KB or 2MB. Therefore divide the total amount of memory required above in MB by the pagesize of 2MB. This is now the number of Hugepages you need to configure.
    351088/2 = 175544
    As user root edit the /etc/sysctl.conf file
    Add/modify vm.nr_hugepages= to be the number of Hugepages calculated.
    vm.nr_hugepages=175544
    Add/modify vm.hugetlb_shm_group = 600
    This parameter is the group id of the TimesTen instance administrator. In the Exalytics system this is oracle. Determine the group id while logged in as oracle with the following command. In this example it’s 600.
    $ id
    $ uid=700(oracle) gid=600(oinstall) groups=600(oinstall),601(dba),700(oracle)
    As user root edit the /etc/security/limits.conf file
    Add/modify the oracle memlock entries so that the fourth field equals the total amount of memory for your TimesTen database. The unit for this value is KB. For example this would be 351088*1024=359514112KB
    oracle hard memlock 359514112
    oracle soft memlock 359514112
    THIS IS VERY IMPORTANT in order for the above changes to take effect you to either shutdown the BI software environment including TimesTen and reboot or issue the following OS command to make the changes permanent.
    $ sysctl -p
    Please note that dynamic setting (including using 'sysctl -p') of vm.nr_hugepages while the system is up may not give you the full number of Hugepages that you have specified. The only guaranteed way to get the full complement of Hugepages is to reboot.
    Check Hugepages has been setup correctly, look for Hugepages_Total
    $ cat /proc/meminfo | grep Huge
    Based on the example values above you would see the following:
    HugePages_Total: 175544
    HugePages_Free: 175544

  • 836: Cannot create data store shared-memory segment, error 22

    Hi,
    I am hoping that there is an active TimesTen user community out there who could help with this, or the TimesTen support team who hopefully monitor this forum.
    I am currently evaluating TimesTen for a global investment organisation. We currently have a large Datawarehouse, where we utilise summary views and query rewrite, but have isolated some data that we would like to store in memory, and then be able to
    report on it through a J2EE website.
    We are evaluating TimesTen versus developing our own custom cache. Obviously, we would like to go with a packaged solution but we need to ensure that there are no limits in relation to maximum size. Looking through the documentation, it appears that the
    only limit on a 64bit system is the actual physical memory on the box. Sounds good, but we want to prove it since we would like to see how the application scales when we store about 30gb (the limit on our UAT environment is 32gb). The ultimate goal is to
    see if we can store about 50-60gb in memory.
    Is this correct? Or are there any caveats in relation to this?
    We have been able to get our Data Store store 8gb of data, but want to increase this. I am assuming that the following error message is due to us not changing the /etc/system on the box:
         836: Cannot create data store shared-memory segment, error 22
         703: Subdaemon connect to data store failed with error TT836
    Can somebody from the User community, or an Oracle Times Ten support person recommend what should be changed above to fully utilise the 32gb of memory, and the 12 processors on the box.
    Its quite a big deal for us to bounce the UAT unix box, so l want to be sure that l have factored in all changes that would ensure the following:
    * Existing Oracle Database instances are not adversely impacted
    * We are able to create a Data Store which is able fully utilise the physical memory on the box
    * We don't need to change these settings for quite some time, and still be able to complete our evaluation
    We are currently in discussion with our in-house Oracle team, but need to complete this process before contacting Oracle directly, but help with the above request would help speed this process up.
    The current /etc/system settings are below, and l have put in the current machines settings as comments at the end of each line.
    Can you please provide the recommended settings to fully utilise the existing 32gb on the box?
    Machine
    ## I have contrasted the minimum prerequisites for TimesTen and then contrasted it with the machine's current settings:
    SunOS uatmachinename 5.9 Generic_118558-11 sun4us sparc FJSV,GPUZC-M
    FJSV,SPARC64-V
    System Configuration: Sun Microsystems sun4us
    Memory size: 32768 Megabytes
    12 processors
    /etc/system
    set rlim_fd_max = 1080                # Not set on the machine
    set rlim_fd_cur=4096               # Not set on the machine
    set rlim_fd_max=4096                # Not set on the machine
    set semsys:seminfo_semmni = 20           # machine has 0x42, Decimal = 66
    set semsys:seminfo_semmsl = 512      # machine has 0x81, Decimal = 129
    set semsys:seminfo_semmns = 10240      # machine has 0x2101, Decimal = 8449
    set semsys:seminfo_semmnu = 10240      # machine has 0x2101, Decimal = 8449
    set shmsys:shminfo_shmseg=12           # machine has 1024
    set shmsys:shminfo_shmmax = 0x20000000     # machine has 8,589,934,590. The hexidecimal translates into 536,870,912
    $ /usr/sbin/sysdef | grep -i sem
    sys/sparcv9/semsys
    sys/semsys
    * IPC Semaphores
    66 semaphore identifiers (SEMMNI)
    8449 semaphores in system (SEMMNS)
    8449 undo structures in system (SEMMNU)
    129 max semaphores per id (SEMMSL)
    100 max operations per semop call (SEMOPM)
    1024 max undo entries per process (SEMUME)
    32767 semaphore maximum value (SEMVMX)
    16384 adjust on exit max value (SEMAEM)

    Hi,
    I work for Oracle in the UK and I manage the TimesTen pre-sales support team for EMEA.
    Your main problem here is that the value for shmsys:shminfo_shmmax in /etc/system is currently set to 8 Gb therby limiting the maximum size of a single shared memory segment (and hence Timesten datastore) to 8 Gb. You need to increase this to a suitable value (maybe 32 Gb in your case). While you are doing that it would be advisable to increase ny of the other kernel parameters that are currently lower than recommended up to the recommended values. There is no harm in increasing them other possibly than a tiny increase in kernel resources, but with 32 GB of RAM I don't think you need be concerned about that...
    You should also be sure that the system has enough swap space configured to supprt a shared memory segment of this size. I would recommend that you have at least 48 GB of swap configured.
    TimesTen should detect that you have a multi-CPU machine and adjust its behaviour accordingly but if you want to be absolutely sure you can set SMPOptLevel=1 in the ODBC settings for the datastore.
    If you want more direct assistance with your evaluation going forward then please let me know and I will contact you directly. Of course, you are free to continue using this forum if you would prefer.
    Regards, Chris

  • Error while activating Data Store Object

    Hi Guru's,
    When I try to activate a data store object i get the error message :
         The creation of the export DataSource failed     
         No authorization to logon as trusted sys tem (Trusted RC=2).     
         No authorization to logon as trusted sys tem (Trusted RC=2).     
         Error when creating the export DataSource and dependent      Program ID 4SYPYCOPQ94IXEGA3739L803Z retrieved for DataStore object ZODS_PRA

    Hi,
    you are facing a issue with your source system 'myself', check and repair it. Also check if the communication user (normally ALEREMOTE) has all permissions needed.
    kind regards
    Siggi

  • Want the data store values to be displayed in input field of form

    Hi,
    I wanted to know wether is there possibility of displaying the data in a input field instead of expression field.
    In our model i have used a form which has material type,plant and Vendor connected to data service1 in turn gives the out put in chart and i also have the second input form which has To(0CALMONTH) combo box which is connected to Data service2 and gives the data in table.
    But what i wanted is to i also want to use the input fields of first form to second Data service2 and get the data based on on inputs of form1 and form2i.e also To field.
    We can connect the port from form1 to DS2 but problem is we need to click the submit of form1 and form2 and it doesnt gives the output according the input of form1 and form2 as it gives the output of that form when we click that submit button.
    I have followed help.sap for data tore procdure.
    So i have used a datstore where i will store the values of form1 and call it in expression field which i will add and hide it in form2(hide because user should not see that input fields).
    Formula used in data store for expression field.
    IF(CONTAINS(STORE@matltype,@Material_Type),STORE@matltype,STORE@matltype &(IF(LEN(STORE@matltype)>0,'; ',''))&@Material_Type)
    But whats happening is the value is getting concatenated when i goon change the values in input field, so i wanted to get the values to be replaced as soon as i change input field of form1(if  use replace function its not working) and also it would be more preferrable if we use input field instead of expression field.
    I would also like to know is there any alternate solution for the above requirement instead of datastore.
    Thank You
    K.Srinivas

    Hi,
          I have Form1 connected to Data service1 displays the data in a Chart and i have another Form2 connected to Dataservice2 displays the data in table.
    In form1 there are Material Type,plant, and Vendor and form2 i have To (0CALMONTH)SO now i want also Form1 inputs for Table which gets teh data from Dataservice2.
    What i have said earlier is connecting Form1 to Dataservice doesnt fetch the data correectly,its because if i click submit in form1 i get the data of those 3 inputs and i need to click the submit button in form2 after giving input which shows the data accordingly where it doesnt fulfill the requirement.
    So i wanted some solution for that.For that reason i have used the data store and the procedure i have followed the help.sap as i said in my above mesage.
    If Data store is also suggestable than i want to display the data in Text input Field  instead of Expression field which should replace the previous values as soon as values change in data store.
    Hope i have tried to be clear if still it is not i am ready to explain again.
    Thank You
    K.Srinivas

Maybe you are looking for

  • Unable to connect to Server  Error in Essbase 11.1.2

    Hi, We are working on Essbase 11.1.2.2. Our applications are in Production and suddenly the  some of the BSO Application are down. We tried to re start the applications but not restarted. So  we restarted the Essbase Hyperion Services in the Server. 

  • Itunes will not play music or other media files suddenly.

    Itunes will be playing a file, usually an audio file, then suddenly it will stop playing mid track. When I tell it to continue using the shortcuts, or the controls, or window bar it says it playing but no volume is produced and the track does not act

  • ITunes.exe

    After installing the ipod update, when you click on itunes the itunes.exe keeps coming up. Itunes will not come up. It just wants to install or uninstall itunes.Any ideas?

  • Cannot start Elements - msg "Elements has stopped working"

    THis is the detail: roblem signature:   Problem Event Name:          APPCRASH   Application Name:          Photoshop Elements 10.0.exe   Application Version:          10.0.0.0   Application Timestamp:          4e70bafc   Fault Module Name:          S

  • Problem with system config "@^@^@^@^@"

    Hello, newbie here (to the forums and to linux in general). I have been trying to install archlinux on my virtual machine for the past few days, following tutorials on youtube and on the archwiki, and I have hit a snag. After installing the packages,