PL/SQL procedure -- log files?
Say when i execute a PL/SQL procedure using SQL* Plus. Is there a place where these executions are stored/logged? Any trace files?
And, when a Java program calls my stored procedure, is there a place these transactions are stored, just to check what exactly is being passed to my stored procedure and what the procedure gave back as result set?
Any pointers will be appreciated. Thanks.
Hi
Use a system.out.println(parametername) to check what values you are passing to the procedure.. you can see the results in your application server console.
Thanks
Similar Messages
-
Hi,
I need to get only a summary log file that show me records by error. for example if the log file has 5000 error for datatype and 300 for PK violation and 2000 for not null value, the problem that the log will show me this error row by row. can I get it summary like this
-5000 rows rejected due to error for datatype
-300 rows rejected due to PK violation
-2000 rows rejected due to not null value
Database: 9i
O/S: Windows 2000 Servereng. Habeeli wrote:
schavali wrote:
Pl see your duplicate post here - big SQL*LDR log file
Pl post what you have found in the documentation so far. AFAIK, there is no way to get a summary log file.
Sriniits not duplicate ,I change to this form that specialized in sql loadre.
I don't found any thing about summary log in the documnetPl post a link to the document you read. Did you see my comment above ?
HTH
Srini -
How to Open Or read SQL Server log file .ldf
Hi all,
How to Open Or read SQL Server log file .ldf
When ever we create database from sql server, it's create two file. (1) .mdf (2) .ldf.
I want to see what's available inside the .ldf file.
Thanks,
AshokI am not too sure but may be the below two undocumented commands might yield the desired result.
DBCC Log
Fn_dblog function
Refer these links for more info,
http://www.mssqlcity.com/Articles/Undoc/SQL2000UndocDBCC.htm
http://blogs.sqlserver.org.au/blogs/greg_linwood/archive/2004/11/27/37.aspx
http://searchsqlserver.techtarget.com/tip/0,289483,sid87_gci1173464,00.html
Some 3rd party tools like Log Explorer can do the job for you.
http://www.lumigent.com/products/le_sql.html
- Deepak -
Problem specifying SQL Loader Log file destination using EM
Good evening,
I am following the example given in the 2 Day DBA document chapter 8 section 16.
In step 5 of 7, EM does not allow me to specify the destination of the SQL Loader log file to be on a mapped network drive.
The question: Does SQL Loader have a limitation that I am not aware of, that prevents placing the log file on a network share or am I getting this error because of something else I am inadvertently doing wrong ?
Note: I have placed the DDL, load file data and steps I follow in EM at the bottom of this post to facilitate reproducing the problem *(drive Z is a mapped drive)*.
Thank you for your help,
John.
DDL (generated using SQL developer, you may want to change the space allocated to be less)
CREATE TABLE "NICK"."PURCHASE_ORDERS"
"PO_NUMBER" NUMBER NOT NULL ENABLE,
"PO_DESCRIPTION" VARCHAR2(200 BYTE),
"PO_DATE" DATE NOT NULL ENABLE,
"PO_VENDOR" NUMBER NOT NULL ENABLE,
"PO_DATE_RECEIVED" DATE,
PRIMARY KEY ("PO_NUMBER") USING INDEX PCTFREE 10 INITRANS 2 MAXTRANS 255 COMPUTE STATISTICS NOCOMPRESS LOGGING TABLESPACE "USERS" ENABLE
SEGMENT CREATION DEFERRED PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 NOCOMPRESS LOGGING STORAGE
INITIAL 67108864
TABLESPACE "USERS" ;
Load.dat file contents
1, Office Equipment, 25-MAY-2006, 1201, 13-JUN-2006
2, Computer System, 18-JUN-2006, 1201, 27-JUN-2006
3, Travel Expense, 26-JUN-2006, 1340, 11-JUL-2006
Steps I am carrying out in EM
log in, select data movement -> Load Data from User Files
Automatically generate control file
(enter host credentials that work on your machine)
continue
Step 1 of 7 ->
Data file is located on your browser machine
"Z:\Documentation\Oracle\2DayDBA\Scripts\Load.dat"
click next
step 2 of 7 ->
Table Name
nick.purchase_orders
click next
step 3 of 7 ->
click next
step 4 of 7 ->
click next
step 5 of 7 ->
Generate log file where logging information is to be stored
Z:\Documentation\Oracle\2DayDBA\Scripts\Load.LOG
Validation Error
Examine and correct the following errors, then retry the operation:
LogFile - The directory does not exist.Hi John,
But, i did'nt found any error when i am going the same what you did.
My Oracle Version is 10.2.0.1 and using Windows xp. See what i did and i got worked
1.I created one table in scott schema :
SCOTT@orcl> CREATE TABLE "PURCHASE_ORDERS"
2 (
3 "PO_NUMBER" NUMBER NOT NULL ENABLE,
4 "PO_DESCRIPTION" VARCHAR2(200 BYTE),
5 "PO_DATE" DATE NOT NULL ENABLE,
6 "PO_VENDOR" NUMBER NOT NULL ENABLE,
7 "PO_DATE_RECEIVED" DATE,
8 PRIMARY KEY ("PO_NUMBER") USING INDEX PCTFREE 10 INITRANS 2 MAXTRANS 255 COMPUTE STATISTICS NOCOMPRESS LOGGING TABLESPACE "USERS" ENABLE
9 )
10 TABLESPACE "USERS";
Table created.I logged into em Maintenance-->Data Movement-->Load Data from User Files-->My Host Credentials
Here i total 3 text boxes :
1.Server Data File : C:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\USERS01.DBF
2.Data File is Located on Your Browser Machine : z:\load.dat <--- Here z:\ means other machine's shared doc folder; and i selected this option (as option button click) and i created the same load.dat as you mentioned.
3.Temporary File Location : C:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\ <--- I did'nt mentioned anything.
Step 2 of 7 Table Name : scott.PURCHASE_ORDERS
Step 3 of 7 I just clicked Next
Step 4 of 7 I just clicked Next
Step 5 of 7 I just clicked Next
Step 6 of 7 I just clicked Next
Step 7 of 7 Here it is Control File Contents:
LOAD DATA
APPEND
INTO TABLE scott.PURCHASE_ORDERS
FIELDS TERMINATED BY ',' OPTIONALLY ENCLOSED BY '"'
PO_NUMBER INTEGER EXTERNAL,
PO_DESCRIPTION CHAR,
PO_DATE DATE,
PO_VENDOR INTEGER EXTERNAL,
PO_DATE_RECEIVED DATE
And i just clicked on submit job.
Now i got all 3 rows in purchase_orders :
SCOTT@orcl> select count(*) from purchase_orders;
COUNT(*)
3So, there is no bug, it worked and please retry if you get any error/issue.
HTH
Girish Sharma -
Logical sql in log file.
Can someone please tell me how to see the complete sql query in the log file. If I run the same query the sql is not being produced I looked in the server log file and also manage sessions log file. It just says all columns from 'Subject Area'. I want to see all the joins and filters as well. Even for repeated queries how can I see complete sql. I set my logging level to 2.
http://lmgtfy.com/?q=obiee+disable+query+caching
http://catb.org/esr/faqs/smart-questions.html#homework -
Repeated Errors in SQL Server log file
I have hundreds of these errors saying 'Login failed for user 'Reporting' The user is not associated with a trusted SQL Server connection [CLIENT: ip address]
The ip address is that of the server that sql server is installed on.
Looking in my log file, all looks good until I get to Service Broker manager has started, then I get Error: 18452, Severity: 14 State: 1 then these two lines repeat about every minute, for the last 3 days!
I think I must have just missed a tick box somewhere, but where?
I have been into one of the databases, and input and checked data, both via an application I wrote and SQL Server Management Studio.
I am also having trouble connecting using my application to connect to the database, I can only connect if I use a Windows administrator account (this SQL Server 2005 running on a Windows 2003 Server, with the app on PC running Windows 2000)Hello Graham,
did you find any solutions for that problem? I'm discovering a similar problem. The state of my error message is 5. According to the following source http://blogs.msdn.com/sql_protocols/archive/2006/02/21/536201.aspx this means the user is not known. This is correct. Neither the SQL-Server nor the Windows system has such a userid.
The faulty user name is 'Reporting'. What is that user used for? I have set up other SQL Server 2005 servers but was never asked for such a name.
Greetings,
Frank -
Hi,
I need to get only a summary log file that show me records by error. for example if the log file has 5000 error for datatype and 300 for PK violation and 2000 for not null value, the problem that the log will show me this error row by row. can I get it summary like this
-5000 rows rejected due to error for datatype
-300 rows rejected due to PK violation
-2000 rows rejected due to not null value
can u give me hand with this?Pl post details of OS and database versions. There is a dedicated SQL*Loader forum at Export/Import/SQL Loader & External Tables
Have you reviewed the documentation for your database version at http://docs.oracle.com ? If so, what have you determined so far ?
HTH
Srini -
We are facing one issue can anyone please help me out.
Logs are getting full in below path . Can we remove the logs ? is it archive logs ?
Path : H:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\DATAIf T-Log backups are configured then the log automatically get truncated once backup is compelted. If you dont have backups configured and you dont want to save the transactions you can manually truncate it
Hi,
From SQL Server 2008 onwards truncate_only is removed, it is replaced by
backup log db_name to disk='Null'
Also if your log is growing AND YOU DONT NEED POINT IN TIME RECOVERY switch recovery model to simple it will force checkpoint and logs will get truncated( If transaction does not requires the log) . Avoid using backup log to null or truncate only.
Swapna,
>>Also if you dont want the transactions to be saved then you can set the recovery model and Simple
This is incorrect language used saving/committing transaction does not depends on recovery model. Recovery model only controls logging and recovery. if transaction is committed it will be present in database no matter what recovery model you use.
Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it.
My TechNet Wiki Articles -
SQL query log File..
Hi All,
I am using a standalone BIP environment... Is there way we can see the actual query log the BIP generates? which location on the server..or how to get it activated.
I want to see what is the actual sql query that is generated along with variables in it.
Fyi : I want to have the query log n not the server error log..
-devHi Eric,
I would like you to re-check on the content level settings here as they are the primary causes of this kind of behavior. You could notice that the same information might have written down in the logical plan of the query too.
Also, as per your description
"In the SOURCES for this logical table, I've set the logical level of the content for E2 appropriately (detail level, same as E1)"
I would like to check on this point again, as if you had mapped E2 to E1 in the same logical source with an inner join, you would get to set the content level at E1 levels themselves but not E2 (Now, that E2 would become a part of the E1 hierarchy too). This might be the reason, the BI Server is choosing to elimiate(null) the values from E2 too (even you could see them in the sql client)
Hope this helps.
Thank you,
Dhar -
Need PL/SQL procedure for file transfer local system to server location
Hi Experts,
The requirement is one file ex: text file,excel file is there in Local system suppose c:newfolder/test.txt.
once run the concurrent program for this particular file in the local system should transfer into server location for the mentioned path.
is it possible to do in PL/SQL or shell script.
can you please share the code.
Thanks
MeherMeher Irk wrote:
The requirement is one file ex: text file,excel file is there in Local system suppose c:newfolder/test.txt.
once run the concurrent program for this particular file in the local system should transfer into server location for the mentioned path.Why do you want to copy a local file from the local client file system, via the Oracle database server, to a remote directory on the server?
Why is the Oracle database server to be used? For what purpose?
There are standard secure methods for a client to copy a file to a server. Such as sftp and scp. No complex Oracle database server layer needed in-between.
So why do you want to use the Oracle database to act as the copy mechanism for you?
It will only make sense if the local file is copied into the database (as a LOB or securefile) and managed by the database. I fail to see why Oracle should be acting as the go-between for the client file system copy process to the server file system. This introduces another moving part in the copy process. Adds more security issues. Adds more authentication issues. And what for? -
SQL Loader Inserting Log File Statistics to a table
Hello.
I'm contemplating how to approach gathering the statistics from the SQL Loader log file to insert them into a table. I've approached this from a Korn Shell Script perspective previously, but now that I'm working in a Windows environment and my peers aren't keen about batch files and scripting I thought I'd attempt to use SQL Loader itself to read the log file and insert one or more records into a table that tracks data uploads. Has anyone created a control file that accomplishes this?
My current environment:
Windows 2003 Server
SQL*Loader: Release 10.2.0.1.0
Thanks,
LukeHello.
Learned a little about inserting into multiple tables with delimited records. Here is my current tested control file:
LOAD DATA
APPEND
INTO TABLE upload_log
WHEN (1:12) = 'SQL*Loader: '
FIELDS TERMINATED BY WHITESPACE
TRAILING NULLCOLS
( upload_log_id RECNUM
, filler_field_0 FILLER
, filler_field_1 FILLER
, filler_field_2 FILLER
, filler_field_3 FILLER
, filler_field_4 FILLER
, filler_field_5 FILLER
, day_of_week
, month
, day_of_month
, time_of_day
, year
, log_started_on "TO_DATE((:month ||' '|| :day_of_month ||' '|| :time_of_day ||' '|| :year), 'Mon DD HH24:MI:SS YYYY')"
INTO TABLE upload_log
WHEN (1:11) = 'Data File: '
FIELDS TERMINATED BY ':'
( upload_log_id RECNUM
, filler_field_0 FILLER POSITION(1)
, input_file_name "TRIM(:input_file_name)"
INTO TABLE upload_log
WHEN (1:6) = 'Table '
FIELDS TERMINATED BY WHITESPACE
( upload_log_id RECNUM
, filler_field_0 FILLER POSITION(1)
, table_name "RTRIM(:table_name, ',')"
INTO TABLE upload_rejects
WHEN (1:7) = 'Record '
FIELDS TERMINATED BY ':'
( upload_rejects_id RECNUM
, record_number POSITION(1) "TO_NUMBER(SUBSTR(:record_number,8,20))"
, reason
INTO TABLE upload_rejects
WHEN (1:4) = 'ORA-'
FIELDS TERMINATED BY ':'
( upload_rejects_id RECNUM
, error_code POSITION(1)
, error_desc
INTO TABLE upload_log
WHEN (1:22) = 'Total logical records '
FIELDS TERMINATED BY WHITESPACE
( upload_log_id RECNUM
, filler_field_0 FILLER POSITION(1)
, filler_field_1 FILLER
, filler_field_2 FILLER
, action "RTRIM(:action, ':')"
, number_of_records
INTO TABLE upload_log
WHEN (1:13) = 'Run began on '
FIELDS TERMINATED BY WHITESPACE
TRAILING NULLCOLS
( upload_log_id RECNUM
, filler_field_0 FILLER POSITION(1)
, filler_field_1 FILLER
, filler_field_2 FILLER
, day_of_week
, month
, day_of_month
, time_of_day
, year
, run_began_on "TO_DATE((:month ||' '|| :day_of_month ||' '|| :time_of_day ||' '|| :year), 'Mon DD HH24:MI:SS YYYY')"
INTO TABLE upload_log
WHEN (1:13) = 'Run ended on '
FIELDS TERMINATED BY WHITESPACE
TRAILING NULLCOLS
( upload_log_id RECNUM
, filler_field_0 FILLER POSITION(1)
, filler_field_1 FILLER
, filler_field_2 FILLER
, day_of_week
, month
, day_of_month
, time_of_day
, year
, run_ended_on "TO_DATE((:month ||' '|| :day_of_month ||' '|| :time_of_day ||' '|| :year), 'Mon DD HH24:MI:SS YYYY')"
INTO TABLE upload_log
WHEN (1:18) = 'Elapsed time was: '
FIELDS TERMINATED BY ':'
( upload_log_id RECNUM
, filler_field_0 FILLER POSITION(1)
, filler_field_1 FILLER
, filler_field_2 FILLER
, elapsed_time
INTO TABLE upload_log
WHEN (1:14) = 'CPU time was: '
FIELDS TERMINATED BY ':'
( upload_log_id RECNUM
, filler_field_0 FILLER POSITION(1)
, filler_field_1 FILLER
, filler_field_2 FILLER
, cpu_time
)Here are the basic table create scripts:
TRUNCATE TABLE upload_log;
DROP TABLE upload_log;
CREATE TABLE upload_log
( upload_log_id INTEGER
, day_of_week VARCHAR2( 3)
, month VARCHAR2( 3)
, day_of_month INTEGER
, time_of_day VARCHAR2( 8)
, year INTEGER
, log_started_on DATE
, input_file_name VARCHAR2(255)
, table_name VARCHAR2( 30)
, action VARCHAR2( 10)
, number_of_records INTEGER
, run_began_on DATE
, run_ended_on DATE
, elapsed_time VARCHAR2( 8)
, cpu_time VARCHAR2( 8)
TRUNCATE TABLE upload_rejects;
DROP TABLE upload_rejects;
CREATE TABLE upload_rejects
( upload_rejects_id INTEGER
, record_number INTEGER
, reason VARCHAR2(255)
, error_code VARCHAR2( 9)
, error_desc VARCHAR2(255)
);Now, if I could only insert a single record to the upload_log table (per table logged); adding separate columns for skipped, read, rejected, discarded quantities. Any advice on how to use SQL Loader to do this (writing a procedure would be fairly simple, but I'd like to perform all of the work in one place if at all possible)?
Thanks,
Luke
Edited by: Luke Mackey on Nov 12, 2009 4:28 PM -
Log files full of service broker errors, but it works OK
We have a C# web application that is using SQL dependency to expire cached query data. Although everything is working okay we are seeing a lot of errors being generated particularly in our production environment.
The first error messages is this:
Service Broker needs to access the master key in the database 'SubscriberManager'. Error code:32. The master key has to exist and the service master key encryption is required.
It shows up in both the server event log and the SQL server log files. I believe the actual content of the messages something of a red herring as I have created a database master key.
I have also tried
Recreating both the service master key and the database master key.
Made sure the database owner is sa.
Made sure the user account has permissions to create services, queues, procedures and subscribe query notifications
Made sure the broker is enabled
I have seen other people with similar errors whilst researching the issue but the error code almost always seems to be 25 or 26 not 32. I have been unable to find anything that tells me what these error codes mean so I'm not sure of the significance.
Also I am seeing a lot of errors like this:
The query notification dialog on conversation handle '{2FA2445B-1667-E311-943C-02C798B618C6}.' closed due to the following error: '-8490Cannot find the remote service 'SqlQueryNotificationService-7303d251-1eb2-4f3a-9e08-d5d17c28b6cf' because
it does not exist.'.
I understand that a certain number of these are normal due to the way that SqlDependency.Stop doesn't clean everything up in the database, but we are seeing thousands of these on one of our production servers.
What is frustrating as we have been using SQL notifications for several years now without issue so something must have changed in either our application or the server setups to cause this but at this point I have no idea what.
The applications are .net 4.0 MVC and WCF running on Windows 2012 servers calling SQL 2012 servers also running on Windows 2012.Hi Mark,
1. for your question about possible memory pressure, if the used memory is below the Max Server Memory, then it's OK. If you have not set Max Server Memory, you should at least leave 4GB for your x64 system.
2. for your original question, I suggest you can check my actions below:
a. run this statement:
Select name, is_master_key_encrypted_by_server
from sys.databases
Where name =
'your_database_name'
if the value of "is_master_key_encrypted_by_server" equals to "0", it means the database does not have a encrypted master key.
b.if there is no encrypted master key, then the error may be hit by the "begin dialog conversation" statement (you can check your sql profiler trace to check".
"Service Broker dialog security lets your application use authentication, authorization, or encryption for an individual dialog conversation (or dialog). By default,
all dialog conversations use dialog security. When you begin a dialog, you can explicitly allow a dialog to proceed without dialog security by including the ENCRYPTION = OFF clause on the BEGIN DIALOG CONVERSATION statement. However, if a remote service binding
exists for the service that the conversation targets, the dialog uses security even when ENCRYPTION = OFF."
(http://msdn.microsoft.com/en-us/library/ms166036.aspx)
Workarounds can be disabling the dialog security (using encryption = off) or create a master key. You can find more information in about URL. -
I am new at the SQL auditing feature. I have played around with it, and I have one problem that I hope someone can help me with. When I first turn on the SQL audit and the audit file is small I can view the log on my local machine with the log viewer in
SSMS. However, as the log files continue to be created and grow, the log file viewer no longer works. It sits there for several minutes saying "initializing log file #1", then just comes back 0 records processed. Sometimes it comes back with an OutOfMemoryException.
I have the SQL audit set to go to files, and the files are set to be 250 MB each.
Is there some better way to look at the SQL audit log files generated, or is there some way to make the log file viewer actually work once the log files begin to grow? I know I can use Transact-SQL statements, but I was hoping to use the log file viewer
or some other GUI viewer. Another issue I'm concerned about is being able to archive the log files to another location and having the ability to look at them.
Thank you.Its looks like bug.
http://connect.microsoft.com/SQLServer/feedback/details/709364/sql-server-audit-logs-do-not-display
http://thomaslarock.com/2012/10/viewing-sql-server-2008-r2-audit-logs-using-ssms-2012/
Use sys.fn_get_audit_file() system function to see the data that you already doing.
http://msdn.microsoft.com/en-us/library/cc280765.aspx
-Prashanth -
Read sql error log, skip lines that are in a exception list
Could someone help, I am creating a nice powershell script to read a sql server log file but to skip lines that are normal in a sql server. the "normal lines are held in a SQL table.
To expand.
I run a query to get the list of exception lines using invoke-sqlcmd this creates $TextEXP
this will contain things like "Microsoft Corporation", "All rights reserved", "Starting up Database"
I then connect to a sql server using SMO and want to read in the error log where the text is not matched the values in $textExpI want to avoid reading extra data and process it. I have a work round but its not a nice clean as its hardcoded the match.
$ENV =$srv.ReadErrorLog()
|? { $_.text
-notmatch'This
is an informational message only'-and$_.text
-notmatch'No
user action is required'-and$_.text
-notmatch'found
0 errors'-and$_.text
-notmatch'Microsoft
Corporation.'-and$_.text
-notmatch'All
rights reserved.'-and$_.text
-notmatch'Server
process ID is'-and$_.text
-notmatch'System
Manufacturer: '-and$_.text
-notmatch'Starting
up database'-and$_.text
-notmatch'Using
''dbghelp.dll'' version'-and$_.text
-notmatch'Authentication
mode is'-and$_.text
-notmatch'Logging
SQL Server messages in file '-and$_.text
-notmatch'Setting
database option'-and$_.text
-notmatch'The
error log has been reinitialized. See the previous log for older entries'-and$_.text
-notmatch'Server
is listening on '-and$_.text
-notmatch'Registry
startup parameters:'-and$_.text
-notmatch'Clearing
tempdb database'-and$_.text
-notmatch'Service
Broker manager has started'-and$_.text
-notmatch'The
Service Broker protocol transport is disabled or not configured'`
-and$_.ProcessInfo
-notmatch"Logon"-and$_.logdate
-ge$Sdate}So after some looking about on the web I found that you can use the | in a string
the following will give an idea of how to use this (this is not a clean bit of code but will give you a starting point)
$TextEXP this is a data table from sql server with the list of values I want to skip
The field name (col name) is extext
Set the string to be empty
$exclusions = ""
#Create a string with the values in $TextExp
Foreach($value in $TextExp){
$exclusions = $exclusions + "$($value.extext)|"
#remove the last pipe from the string
$exclusions = $exclusions.substring(0,$exclusions.length-1)
##This will create a long string value|value|value###
$err = $srv.readerrorLog() | ?{$_.text - notmatch $exclusions}
###end
May need bit of a clean up and may be a better way but seems to do what I need for now.
Thanks all for the help -
Display data in log file using PL/SQL procedure
Just as srw.message is used in Oracle RDF Reports to display data in log file in Oracle Apps, similarly how it is possible to display data in log file using PL/SQL procedure?
Please also mention the syntax too.Pl post details of OS, database and EBS versions.
You will need to invoke the seeded FND_LOG procedure - see previous discussions on this topic
Enable debug for pl/sql
https://forums.oracle.com/forums/search.jspa?threadID=&q=FND_LOG&objID=c3&dateRange=all&userID=&numResults=15&rankBy=10001
HTH
Srini
Maybe you are looking for
-
Issue to add t-codes & functions in ERM role request GRC 10.0 - SP 5
Hi We are on SAP GRC 10.0 - SP5. We are facing a strange issue while creating a Role request. After Defining the role, we are unable to add actions and there is no option of add/delete. Similarly while adding Functions, we are getting an error of Fun
-
I just upgraded to 6.0.4.2 and now my purchased list is all messed up
Any one know how to fix this? I just up graded to 6.0.4.2 Now when I look at my purchased list of songs it shows a bunch of old songs that I never purchased through ITunes? and many of the songs that I did purchase are showing up in the list 4 and 5
-
Is Online number available for use in Thailand? If not, when will it be available? Thanks in advance for your response. Solved! Go to Solution.
-
IMac G5 iSight Model. NEW HDD.
Hi. I just replaced the old HDD with my New 500 GB Seagate Barracuda HDD. So, I'm trying to install new OS X Tiger too. When I put the disc in and turn it on with holding down the C key... It goes to the grey screen with apple sign on it and then sta
-
Does anyone know if there is an additional charge to use the LR Mobile app on your iPad if you are a CC Photography member? Most things seem to suggest it is part of the package but on the CC LR web site the following note is included: Note: If you