What if all the redo logs of a database are lost
I want to know how a database can be recovered if all the three redo logs are lost at the same time.
Thanks,
Prabhath.
You will be able to find the detail procedures of this kind of recover here:
Backup and Recovery Concepts Contents / Search / Index / PDF
Backup and Recovery Documentation Online Roadmap Contents / Search / /
Recovery Manager Quick Reference Contents / Search / / PDF
Recovery Manager Reference Contents / Search / Index / PDF
Recovery Manager User's Guide Contents / Search / Index / PDF
http://otn.oracle.com/pls/db92/db92.docindex?remark=homepage
if you are making backups to the database using RMAN the recover is easier.
Joel Pérez
Similar Messages
-
Redo logs and Flash recovery area
Hi,
Is it a good practice to place a copy of the (multiplexed) online redo at the flash recovery area? Wouldn't it be better to place a copy of the archived log at the flash recovery area?user492400 wrote:
Hi,
Is it a good practice to place a copy of the (multiplexed) online redo at the flash recovery area? Wouldn't it be better to place a copy of the archived log at the flash recovery area?Its not only the archvielogs that should be placed in the FRA. FRA is supposed to contain one copy of the archive logs and the rest 9 destinations are given to you for the multiplexing of it. The idea of multiplexing the redo logs and placing them anywhere( not just on the FRA itself) is simply required so that you won't get to a situation where you would lose all the redo log files and thus have to recreate them, losing the data inside them. So aleast one copy of the log files should be there and where you want to put it, that would depend on you.
HTH
Aman.... -
Where all the redo/archive log volume come from?
Hi,
I have a database with a size of around 50G. Thats not very much, but the crazy thing is that this database has a redo log volume per day that is nearly equal the size of the database:
-- Daily View
SQL> SELECT TRUNC(completion_time) AS time, ROUND(SUM(blocks * block_size)/1024/1024,2) AS size_mb, ROUND(SUM(blocks * block_size)/1024/1024/1024,2) AS size_gb FROM v$archived_log GROUP BY TRUNC(completion_time) ORDER BY 1 DESC;
TIME SIZE_MB SIZE_GB
2013-02-03 49599.7 48.44
2013-02-02 50104.63 48.93
2013-02-01 54466.36 53.19
2013-01-31 52935.41 51.69
2013-01-30 51596.85 50.39
-- Hourly View
SQL> SELECT TRUNC(completion_time,'HH24') AS time, ROUND(SUM(blocks * block_size)/1024/1024,2) AS size_mb, ROUND(SUM(blocks * block_size)/1024/1024/1024,2) AS size_gb FROM v$archived_log GROUP BY TRUNC(completion_time,'HH24') ORDER BY 1 DESC;
TIME SIZE_MB SIZE_GB
2013-02-03 23 160.68 .16
2013-02-03 22 163.73 .16
2013-02-03 21 195.65 .19
2013-02-03 20 4492.44 4.39
2013-02-03 19 176.3 .17
2013-02-03 18 4259.8 4.16
2013-02-03 17 226.99 .22
2013-02-03 16 5465.64 5.34
2013-02-03 15 166.47 .16
2013-02-03 14 165.05 .16
2013-02-03 13 253.33 .25
2013-02-03 12 9568.93 9.34
2013-02-03 11 391.14 .38
2013-02-03 10 9585.36 9.36
2013-02-03 09 160.02 .16
2013-02-03 08 200.1 .2
2013-02-03 07 166.66 .16
2013-02-03 06 267.45 .26
2013-02-03 05 309.26 .3
2013-02-03 04 1486.87 1.45
2013-02-03 03 11106.77 10.85
2013-02-03 02 318.1 .31
2013-02-03 01 147.47 .14
2013-02-03 00 165.47 .16
2013-02-02 23 160.17 .16
2013-02-02 22 159.25 .16
2013-02-02 21 205.9 .2
2013-02-02 20 4473.27 4.37
2013-02-02 19 124.64 .12
2013-02-02 18 4364.69 4.26
2013-02-02 17 295.02 .29
2013-02-02 16 5415.06 5.29
2013-02-02 15 166.22 .16
2013-02-02 14 166.83 .16
2013-02-02 13 84.62 .08
2013-02-02 12 9905.55 9.67
2013-02-02 11 418.31 .41
2013-02-02 10 9548.78 9.32
2013-02-02 09 166.51 .16
2013-02-02 08 204.88 .2
2013-02-02 07 167.37 .16I know that every database change produces redo, but why that much?
Is there a way to investigate further where all the redo information comes from?
Every 2 hours redo is backed up to tape.
There are no significant batches, shaking and shuffling the data around all the time.
Nearly all changes in this database is happening through user inputs.
Database is 11gR2 on Redhat 5.6.
Thanks
941743
Edited by: 941743 on 04.02.2013 15:32
Edited by: 941743 on 04.02.2013 15:33I am afraid there is no easy and direct way to pull the answer from the data dictionary.
But with some work, you can get to the bottom of this.
This query shows you how much redo is used by the sessions in your system currently:
select a.sid , a.value from v$sesstat a, v$statname b
where a.statistic# = b.statistic#
and b.name = 'redo size'Unfortunately, I am not aware of a AWR view (license required!) that corresponds to v$sesstat, so it would not be easy to look back and see which session generated the redo.
One way is to troubleshoot the problem online, using a utility such as snapper- http://blog.tanelpoder.com/2010/03/22/oracle-session-snapper-v3-10/ . See which sessions generate most redo and see what SQL they are running at that time.
The other way is to create your own “repository”. Sample V$SESSION, v$sesstat , v$statname and possibly V$SQL frequently (every few seconds) and record how much redo and what SQLs are running by each session. Once you gather enough data, look which sessions had the biggest “redo size” delta and what SQL were they running at that time.
Iordan Iotzov
http://iiotzov.wordpress.com/ -
Bottleneck when switching the redo log files.
Hello All,
I am using Oracle 11.2.0.3.
The application team reported that they are facing slowness at certain time.
I monitored the database and I found that at some switching of the redo log files (not always) I am facing a slowness at the application level.
I have 2 threads since my database is RAC, each thread have 3 redo log groups multiplexed to the FRA, with size 300 MB each.
Is there any way to optimize the switch of redo log files? knowing that my database is running in ARCHIVELOG mode.
Regards,Hello Nikolay,
Thanks for your input I am sharing with you the below information. I have 2 instances so I will provide the info from each instance
Instance 1:
Load Profile Per Second Per Transaction Per Exec Per Call
~~~~~~~~~~~~ --------------- --------------- ---------- ----------
DB Time(s): 4.9 0.0 0.00 0.00
DB CPU(s): 1.1 0.0 0.00 0.00
Redo size: 3,014,876.2 3,660.4
Logical reads: 32,619.3 39.6
Block changes: 7,969.0 9.7
Physical reads: 0.2 0.0
Physical writes: 164.0 0.2
User calls: 7,955.4 9.7
Parses: 288.9 0.4
Hard parses: 96.0 0.1
W/A MB processed: 0.2 0.0
Logons: 0.9 0.0
Executes: 2,909.4 3.5
Rollbacks: 0.0 0.0
Instance 2:
Load Profile Per Second Per Transaction Per Exec Per Call
~~~~~~~~~~~~ --------------- --------------- ---------- ----------
DB Time(s): 5.5 0.0 0.00 0.00
DB CPU(s): 1.4 0.0 0.00 0.00
Redo size: 3,527,737.9 3,705.7
Logical reads: 29,916.5 31.4
Block changes: 8,893.7 9.3
Physical reads: 0.2 0.0
Physical writes: 194.0 0.2
User calls: 7,742.8 8.1
Parses: 262.7 0.3
Hard parses: 99.5 0.1
W/A MB processed: 0.4 0.0
Logons: 1.0 0.0
Executes: 2,822.5 3.0
Rollbacks: 0.0 0.0
Transactions: 952.0
Instance 1:
Top 5 Timed Foreground Events
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Avg
wait % DB
Event Waits Time(s) (ms) time Wait Class
DB CPU 1,043 21.5
log file sync 815,334 915 1 18.9 Commit
gc buffer busy acquire 323,759 600 2 12.4 Cluster
gc current block busy 215,132 585 3 12.1 Cluster
enq: TX - row lock contention 23,284 264 11 5.5 Applicatio
Instance 2:
Top 5 Timed Foreground Events
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Avg
wait % DB
Event Waits Time(s) (ms) time Wait Class
DB CPU 1,340 24.9
log file sync 942,962 1,125 1 20.9 Commit
gc buffer busy acquire 377,812 594 2 11.0 Cluster
gc current block busy 211,270 488 2 9.1 Cluster
enq: TX - row lock contention 30,094 299 10 5.5 Applicatio
Instance 1:
Operating System Statistics Snaps: 1016-1017
-> *TIME statistic values are diffed.
All others display actual values. End Value is displayed if different
-> ordered by statistic type (CPU Use, Virtual Memory, Hardware Config), Name
Statistic Value End Value
AVG_BUSY_TIME 17,451
AVG_IDLE_TIME 81,268
AVG_IOWAIT_TIME 1
AVG_SYS_TIME 6,854
AVG_USER_TIME 10,548
BUSY_TIME 420,031
IDLE_TIME 1,951,741
IOWAIT_TIME 288
SYS_TIME 165,709
USER_TIME 254,322
LOAD 3 6
OS_CPU_WAIT_TIME 523,000
RSRC_MGR_CPU_WAIT_TIME 0
VM_IN_BYTES 311,280
VM_OUT_BYTES 75,862,008
PHYSICAL_MEMORY_BYTES 62,813,896,704
NUM_CPUS 24
NUM_CPU_CORES 6
NUM_LCPUS 24
NUM_VCPUS 6
GLOBAL_RECEIVE_SIZE_MAX 4,194,304
GLOBAL_SEND_SIZE_MAX 4,194,304
TCP_RECEIVE_SIZE_DEFAULT 16,384
TCP_RECEIVE_SIZE_MAX 9.2233720368547758E+18
TCP_RECEIVE_SIZE_MIN 4,096
TCP_SEND_SIZE_DEFAULT 16,384
TCP_SEND_SIZE_MAX 9.2233720368547758E+18
TCP_SEND_SIZE_MIN 4,096
Operating System Statistics - Detail Snaps: 1016-101
Snap Time Load %busy %user %sys %idle %iowait
22-Aug 11:33:55 2.7 N/A N/A N/A N/A N/A
22-Aug 11:50:23 6.2 17.7 10.7 7.0 82.3 0.0
Instance 2:
Operating System Statistics Snaps: 1016-1017
-> *TIME statistic values are diffed.
All others display actual values. End Value is displayed if different
-> ordered by statistic type (CPU Use, Virtual Memory, Hardware Config), Name
Statistic Value End Value
AVG_BUSY_TIME 11,823
AVG_IDLE_TIME 86,923
AVG_IOWAIT_TIME 0
AVG_SYS_TIME 4,791
AVG_USER_TIME 6,991
BUSY_TIME 475,210
IDLE_TIME 3,479,382
IOWAIT_TIME 410
SYS_TIME 193,602
USER_TIME 281,608
LOAD 3 6
OS_CPU_WAIT_TIME 615,400
RSRC_MGR_CPU_WAIT_TIME 0
VM_IN_BYTES 16,360
VM_OUT_BYTES 72,699,920
PHYSICAL_MEMORY_BYTES 62,813,896,704
NUM_CPUS 40
NUM_CPU_CORES 10
NUM_LCPUS 40
NUM_VCPUS 10
GLOBAL_RECEIVE_SIZE_MAX 4,194,304
GLOBAL_SEND_SIZE_MAX 4,194,304
TCP_RECEIVE_SIZE_DEFAULT 16,384
TCP_RECEIVE_SIZE_MAX 9.2233720368547758E+18
TCP_RECEIVE_SIZE_MIN 4,096
TCP_SEND_SIZE_DEFAULT 16,384
TCP_SEND_SIZE_MAX 9.2233720368547758E+18
TCP_SEND_SIZE_MIN 4,096
Operating System Statistics - Detail Snaps: 1016-101
Snap Time Load %busy %user %sys %idle %iowait
22-Aug 11:33:55 2.6 N/A N/A N/A N/A N/A
22-Aug 11:50:23 5.6 12.0 7.1 4.9 88.0 0.0
------------------------------------------------------------- -
What will heppen if redo logs at os level get deleted
Friends,
need 1 answer about the query: what will heppen if redo logs at os level get deleted.how can i find where are my multiplexed redo logs, so that i can copy it from there
SQL> shutdown abort
ORACLE instance shut down.
SQL> startup mount
ORACLE instance started.
Total System Global Area 534462464 bytes
Fixed Size 2256912 bytes
Variable Size 398462960 bytes
Database Buffers 125829120 bytes
Redo Buffers 7913472 bytes
Database mounted.
SQL> select * from v$logfile
2 ;
GROUP# STATUS TYPE
MEMBER
IS_
3 ONLINE
C:\O\ORADATA\XE\REDO03.LOG
NO
2 ONLINE
C:\O\ORADATA\XE\REDO02.LOG
NO
GROUP# STATUS TYPE
MEMBER
IS_
1 ONLINE
C:\O\ORADATA\XE\REDO01.LOG
NO However, Oracle knows about the multiplexing, so if your logs were multiplexed properly (using redo log members), Oracle would already have known about them and identified that there was one available.
Just goes to show you (if this is, in fact what you did), don't just go blindly deleting files that have a ".LOG" extension. -
How to change the redo log file location.... ?
I want all my redo log files to be created in \u10 instead of current /u01?
How to do it? NOARCHIVELOG mode database on Oracle 10g R2.
Thank you,
SmirhHi..
I want all my redo log files to be created in \u10 instead of current /u01?I think it should be /u10 :)...
Anand
Edited by: Anand... on Nov 5, 2009 5:32 AM Removed the misinformation about downtime -
HT201359 how can i find out what all the charges on my checking account are for?
how can i find out what all the charges on my checking account are for?
If you log into your account on your computer's iTunes via the Store > View Account menu option, you should then see a Purchase History section with a 'see all' link to the right of it. Click on that and you should then see a list of your purchases - the most recent purchase is above the list. I'm not sure if that shows auto-renewing subscriptions (I don't have any so I can't check).
If you can't find the charges and you haven't added or changed your card details on your account (when you do then a small temporary store holding charge may be applied to check that the details are correct and valid, which should disappear within a few days) then you can contact iTunes Support via this page : http://www.apple.com/support/itunes/contact/ - click on Contact iTunes Store Support on the right-hand side of the page, then Purchases, Billing & Redemption -
GoldenGate : Why to read all the Redo or Archive, can't we skip ??
Hi All,
I have a big doubt if anyone can give clarity. The scenario is :- In my environment data is flowing from Oracle to Oracle but Huge transactions in Millions and we use this tool basically to improve LAG when we get data for reporting.
Now let's say I have database user A which is doing some temporary transaction on source for couple of hours but it is HUGE and generating lot of archives AND we don't want to flow any data from the transaction being performed by user A - so we use EXCLUDEUSER A parameter.
Now during same time if any other user say B is performing some operation in one table - we want it to flow, So it is flowing normally BUT the main problem is LAG increased like anything cause GoldenGate is now reading all the redo or archives which of no use and data is flowing very slowly :( :(
My point was will it be also possible that I don't want my GoldenGate to read any redo log or archives if the user A is performing operation and read only those archives if user B or any other user is performing some operation. This will improve the LAG cause even though we ignore the particular user, It is still going to read all the redo or huge archives generated by user A or user B or whatever in sequence, Please correct me if I am wrong ?
I suppose goldengate is going to read all the transaction regardless of A or B user. Only it will ignore the A and process transaction of B.
I know it reads in sequence but can't it read specific archives which is useful instead of reading all.Hi Steven,
Thanks for your reply.
I agree with you that "normal" mode of GoldenGate is to capture changes and replicate them and Ofcourse I want it to capture any changes but I don't want it to pass through reading all the archives if I am ignoring one user who is performing HUGE temporary transaction and generating huge archives.
It is going to slow down all the other small transaction changes cause it is passing through all those archives also which is of NO use. Do you mean that "Integrated mode" of new Goldengate version will work in that case ?
Please help, I really need this information. It will help everyone. -
What will happen when redo log file or archive log file, which is yet to be
What will happen when redo log file or archive log file, which is yet to be read by logminer is corrupted? It seems that the captures process hangs between “Paused for flow control” and “Enqueuing Messages”. How to come out of this condition without recreating the captures process?
Any clue is helpful
Thanks in advance for your help.Basically you can't skip SCN since it will result in a data integrity issues (say you had skipped some inserts and later there will be some updates to a not replicated data).
Streams maintain their own checkpoint tables with transaction related stuff. So there is no way you can jump over a range of SCN's without recreating capture.
The only thing you can try - temporary give capture process a rule without any objects. But it will need to mine through the redo anyway. -
What r all the tables used for this report. please reply
hai,
what r all the tables used for this report.
report :
<b>
Report to display all finished goods that go out-of-stock. Developed a drill down report for materials details (totals and subtotals for material stock values by material group and material type).</b>
thanks in advancehi Ashok,
Check this out
http://www.allsaplinks.com/tables.html
http://www.sapgenie.com/abap/tables.htm
Regards,
Santosh -
What r all the tables used for the following report
hai
what r all the tables used for the following report
report:
<b>
Report that displays all the late shipments in a particular period of time</b>
send important fields for this tables also
thanks in advance.Hi,
Imp table are vttk,vttp,vtts.
Regards
Amole -
Sizing the redo log files using optimal_logfile_size view.
Regards
I have a specific question regarding logfile size. I have deployed a test database and i was exploring certain aspects with regards to selecting optimal size of redo logs for performance tuning using optimal_logfile_size view from v$instance_recovery. My main goal is to reduce the redo bytes required for instance recovery. Currently i have not been able to optimize the redo log file size. Here are the steps i followed:-
In order to use the advisory from v$instance_recovery i had to set fast_start_mttr_target parameter which is by default not set so i did these steps:-
1)SQL> sho parameter fast_start_mttr_target;
NAME TYPE VALUE
fast_start_mttr_target integer 0
2) Setting the fast_start_mttr_target requires nullifying following deferred parameters :-
SQL> show parameter log_checkpoint;
NAME TYPE VALUE
log_checkpoint_interval integer 0
log_checkpoint_timeout integer 1800
log_checkpoints_to_alert boolean FALSE
SQL> select ISSES_MODIFIABLE,ISSYS_MODIFIABLE,ISINSTANCE_MODIFIABLE,ISMODIFIED from v$parameter where name like'log_checkpoint_timeout';
ISSES_MODIFIABL ISSYS_MODIFIABLE ISINSTANCE_MODI ISMODIFIED
FALSE IMMEDIATE TRUE FALSE
SQL> alter system set log_checkpoint_timeout=0 scope=both;
System altered.
SQL> show parameter log_checkpoint_timeout;
NAME TYPE VALUE
log_checkpoint_timeout integer 0
3) Now setting fast_start_mttr_target
SQL> select ISSES_MODIFIABLE,ISSYS_MODIFIABLE,ISINSTANCE_MODIFIABLE,ISMODIFIED from v$parameter where name like'fast_start_mttr_target';
ISSES_MODIFIABL ISSYS_MODIFIABLE ISINSTANCE_MODI ISMODIFIED
FALSE IMMEDIATE TRUE FALSE
Setting the fast_mttr_target to 1200 = 20 minutes of checkpoint switching according to Oracle recommendation
Querying the v$instance_recovery view
4) SQL> select ACTUAL_REDO_BLKS,TARGET_REDO_BLKS,TARGET_MTTR,ESTIMATED_MTTR, OPTIMAL_LOGFILE_SIZE,CKPT_BLOCK_WRITES from v$instance_recovery;
ACTUAL_REDO_BLKS TARGET_REDO_BLKS TARGET_MTTR ESTIMATED_MTTR OPTIMAL_LOGFILE_SIZE CKPT_BLOCK_WRITES
276 165888 *93* 59 361 16040
Here Target Mttr was 93 so i set the fast_mttr_target to 120
SQL> alter system set fast_start_mttr_target=120 scope=both;
System altered.
Now the logfile size suggested by v$instance_recovery is 290 Mb
SQL> select ACTUAL_REDO_BLKS,TARGET_REDO_BLKS,TARGET_MTTR,ESTIMATED_MTTR, OPTIMAL_LOGFILE_SIZE,CKPT_BLOCK_WRITES from v$instance_recovery;
ACTUAL_REDO_BLKS TARGET_REDO_BLKS TARGET_MTTR ESTIMATED_MTTR OPTIMAL_LOGFILE_SIZE CKPT_BLOCK_WRITES
59 165888 93 59 290 16080
After altering the logfile size to 290 as show below by v$log view :-
SQL> select GROUP#,THREAD#,SEQUENCE#,BYTES from v$log;
GROUP# THREAD# SEQUENCE# BYTES
1 1 24 304087040
2 1 0 304087040
3 1 0 304087040
4 1 0 304087040
5 ) After altering the size i have observed the anomaly as redo log blocks to be applied for recovery has increased from *59 to 696* also now v$instance_recovery view is now suggesting the logfile size of *276 mb*. Have i misunderstood something
SQL> select ACTUAL_REDO_BLKS,TARGET_REDO_BLKS,TARGET_MTTR,ESTIMATED_MTTR, OPTIMAL_LOGFILE_SIZE,CKPT_BLOCK_WRITES from v$instance_recovery;
ACTUAL_REDO_BLKS TARGET_REDO_BLKS TARGET_MTTR ESTIMATED_MTTR OPTIMAL_LOGFILE_SIZE CKPT_BLOCK_WRITES
*696* 646947 120 59 *276* 18474
Please clarify the above output i am unable to optimize the logfile size and have not been able to achieve the goal of reducing the redo log blocks to be applied for recovery, any help is appreciated in this regard.sunny_123 wrote:
Sir oracle says that fast_start_mttr target can be set to 3600 = 1hour. As suggested by following oracle document
http://docs.oracle.com/cd/B10500_01/server.920/a96533/instreco.htm
I set mine value to 1200 = 20 minutes. Later i adjusted it to 120=2 minutes as Target_mttr suggested it to be around 100 (if fast_mttr_target value is too high or too low effective value is contained in target_mttr of v$instance_recovery)Just to add, you are reading the documentation of 9.2 and a lot has changed since then. For example, in 9.2 the parameter FSMTTR was introduced and explicitly required to be set and monitored by the DBA for teh additional checkpoint writes which might get caused by it. Since 10g onwards this parameter has been made automatically maintained by Oracle. Also it's been long that 9i has been desupported followed by 10g so it's better that you start reading the latest documentation of 11g and if not that, at least of 10.2.
Aman.... -
I know OS X Server uses PostgreSQL now instead of MySQL which is fine by me. I am trying to figure out what is in the PostgreSQL_Server_Services.log file in Library/Logs/PostgreSQL/ . It is about 188 GB right now and ideally instead of relocating it along with other services to another drive from the nice 256GB SSD that is in the system I would like to understand what is in the database and how to manage it's size better.
Any pointers for best practice on managing the database or what the heck is in it?
Thanks!Thanks for the response. I mistyped in the question as I understood it was a log file. The DB itself is only about 1-2GB. As for reviewing I am fine with modifying the permissions and examining the file in the console or other app but I am most interested in how to manage it safely. Can I just clear the log contents? If so what is the safe way to do so? I know with many DB's the log file is critical to it's function and if things happen to the log file it can render the DB unusable. In addition what is the best way to modify the rotation routine it uses and set the logging level? (I have pasted the .plist contents for Postgresql for Server Services below)
I have also downloaded and installed pgAdmin but have not gone through the steps of connecting it to the DB and log file (users setup etc.)
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE plist PUBLIC "-//Apple Computer//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
<plist version="1.0">
<dict>
<key>ProgramArguments</key>
<array>
<string>-D</string>
<string>/Library/Server/PostgreSQL For Server Services/Data</string>
<string>-c</string>
<string>unix_socket_directory=/Library/Server/PostgreSQL For Server Services/Socket</string>
<string>-c</string>
<string>logging_collector=on</string>
<string>-c</string>
<string>log_connections=on</string>
<string>-c</string>
<string>log_lock_waits=on</string>
<string>-c</string>
<string>log_statement=ddl</string>
<string>-c</string>
<string>log_line_prefix=%t </string>
<string>-c</string>
<string>listen_addresses=</string>
<string>-c</string>
<string>log_directory=/Library/Logs/PostgreSQL</string>
<string>-c</string>
<string>log_filename=PostgreSQL_Server_Services.log</string>
<string>-c</string>
<string>unix_socket_group=_postgres</string>
<string>-c</string>
<string>unix_socket_permissions=0770</string>
</array>
</dict>
</plist> -
What's all the fuss about iOS6 maps???
For starters, I'm not an iFan... My first smartphone was a BB and I used to swear by it. I found (and sometimes still do) that the iPhone was more of a toy while the BB was a professional tool. I still prefer the BB Bold's keyboard for typing emails than the iPhone's touchscreen keyboard. I ended up buying a 4S earlier this year (March) only because BB had no Outlook Exchange license available where I lived (Maldives) and it was either Android or iPhone, so the choice was clear there. Since then, I've come to grow quite fond of it and am quite happy with the integrations and functionality of the phone albeit it being slightly more time-consuming to send out emails.
The iPhone 5 launch event came out and I looked at all the details. Here's my conclusion: an extra row or icons, a slightly faster processor and most of all 4G. Now, I don't really need an extra row of icons... In fact, I sometimes find the 4S a little too long (jeans pockets along with a pack of cigarettes: cigarette pack usually loses and looks like it's been trampled by a herd of elephants...). So, I'm not in for a bigger screen as the current one's just fine. I never found my 4S to lack speed so I don't really care about the faster processor either. 4G? I love it! But I live in China now... It will be a while before it'll come out here, so why bother for the moment? It's complicated enough as it is as China Mobile (biggest network and best coverage, which I am also currently on) doesn't even have 3G for the iPhone. Only EDGE, but I make do... China Unicom offer 3G for iPhones but not as good a coverage... So, what to do? So in the end, a 4S with the brand new iOS6 update to me sounds better than going out to buy a new iPhone 5.
Update has been out for 3 days now. And before I went ahead and updated I checked the reviews, comments and discussions... WOW! It would seem Apple has opened the wraths of **** upon the world and that the Apocalypse is upon us when reading on about all the bugs and issues. And everyone goes on and on about the maps... It's made it onto CNN, BBC and just about anyone with an iota of grey matter and a keyboard. There are even comments mentioning that Asia is practically mapless with the new native app. I froze! No way in **** was I going to update if I was going to lose such functionality. I've just arrived here 2 months ago, don't speak a word of the local language, feel like getting an audience with the Pope is simpler than giving a cab driver directions and the Google Maps app has often been a lifesaver for me here so far... As much as I want to get my hands on the rest of the new functions, it was a risk not worth taking.
Today I bump into a colleague of mine at work and he's just updated his iPad (2nd gen) to iOS6. I tell him all I've heard about the maps and he brings it up to the restaurant for lunch. I'm scrolling through the new map app and honestly don't find much of a difference. Now, seeing as everyone is whining about the uselessness of the directions the new app gives, I give it a test. I search for the Park Hyatt in Beijing. The Google Maps app used to give me 3 different results. The new app gives me one result only, it's correct and gives me the right directions to get there...
So, what is all the fuss on the new update??? So Apple got rid of Google Maps... Yes, it'll take time to get better but Apple so far have usually managed to do a pretty good job at doing that. At first, everyone was complaining about battery drainage... Nobody mentions it anymore... Strange... Is this intentional Apple bashing?... From what I've seen so far, I still don't get what all the fuss is about.
I'm going ahead with the update... Will let you know how it goes!Hi,
What's the worst that I've spent 100 Eur alltogether for GPS map based applications for several sports - golf, bike, ski, etc. These applications use data from iPhone Map application and its disaster after my update onto iOS 6. These are pictures of 2 of my favorit golf resorts:
And here one example from a place where I use to bike:
Very nice, isn't it?
And unfortunately this problem won't disappear even when Google put stand alone map application into iStore.
So I have only way to downgrade to 5.1.1 immediately to not throw away money I already spent.
Ivan -
I use iCloud. I have an iPhone, Apple laptop, and I recently bought an iPad. I can view the pictures I take on my cell phone on my iPad. But what about all the pictures I took on my phone before I bought my iPad? How do I view those on my iPad?
POP: yahoo, aol, comcast/time warnder/road runner
Imap: google, hotmail, and more including icloud.
If you want to use multiple devices - move to imap, I would even say "exchange", but Google does not support free exchange anymore, since January of that year. So Ironically Icloud or Hotmail would be my choices right now.
To find out more about what happens to you, search on Google "difference between pop and imap"
Maybe you are looking for
-
How do I get a data plan on my ipad.
My friend bought an ipad 2 with wifi and 3G, and used to have a data plan until she switched it to her new ipad. I would like to buy a plan, with AT&T, but I have no idea how. Can someone helps me?
-
Opening .pdf files in Firefox 3.6.13 spawns "infinite" tabs -- amazing!
If I click a link to a .pdf file, Firefox will open up tab after tab after tab -- they go shooting across the top of the tab bar (it's incredible). I can use Task Manager to force FF to quit (and then manually end the Firefox.exe process since it doe
-
ESB: JGroups channel error
Hi All, Recently i hit the following error when i tried to register the ESB CustomerData example using endpoint activations: "JGroups channel error" The error stack shows also: org.collaxa.thirdparty.jgroups.ChannelException: failed to start protocol
-
calculates and returns the sum of all the odd numbers from 1 to x.? this is what i have so far: for (count = 1; i = x; i++) sum = 0; count + 2; sum = count i = i + 2 when i compile, i get an error because count + 2; is not a statement. why? Your help
-
SRM BBP_BACKEND_DEST Syst_type of CATALOG
Hi, We noticed an entry in our SRM 5.0 BBP_BACKEND_DEST table for syst type CATALOG. The logical system name pointed to an SRM RFC destination (to it self) and it was also defined as a Local system type. What is this entry used for and should it be d