Best way to handle single line Customer and Address Data

Hi, We have customer data in a single line means name, address everything is in a single line. I am now suppose to cleanse this data using Data Quality. Now my question is how to parse this data. Normally I have dealt with multiple line or address data which I can map to input fields of the Address cleanse transforms. In this case, Shall I just map that single field to "Address_line" of the Address cleanse transform? It should be able to parse or shall I parse all the fields using some function in Query transform and then feed into Address Cleanse transform?
Thanks,

Hi, We have customer data in a single line means name, address everything is in a single line. I am now suppose to cleanse this data using Data Quality. Now my question is how to parse this data. Normally I have dealt with multiple line or address data which I can map to input fields of the Address cleanse transforms. In this case, Shall I just map that single field to "Address_line" of the Address cleanse transform? It should be able to parse or shall I parse all the fields using some function in Query transform and then feed into Address Cleanse transform?
Thanks,

Similar Messages

  • Best way to handle inversion of custom fields?

    We've added some additional numeric fields, such as quantities, to some of the LO extractors.
    What is the best way to handle the reversal of these additional fields when reversal records come over on the extractor?  I tried switching on the "inversion" switch in the extractor and this didn't work - found OSS note 382779 which explains that this is expected behavior, that one must put some logic in the user exit to do the inversion.
    Is it just a matter of coding if the ROCANCEL field (mapped to 0STORNO in BW) is 'X' or 'R' to flip the sign of the numeric field we added?  But is it better to put this in the R/3 user exit or in a start routine in the transfer rules do you think?  Does flipping on the "inversion" switch make any difference at all in this case?
    Thanks for your help!
    Chris

    Hey,
       I dont this inversion which matters  as you said you added quantity type of fields. If i were you i will do the user exit which can take care good but tranfer routine also helps in most of the cases.

  • Best way to set up a custom table using dates ytd, quarters, months

    Hello-
    I did post this on the crystal forum however it really involves setting up a good structured table in order to get the data to report on which I think we need to change which is why I'm posting here.
    I am not a dba but I work with crystal reports and we are working together to get data in tables that we can report on.  We are in the process of creating a data warehouse, which will mainly be summarized data we are exporting out of our legacy system and importing into a mysql database.  Most of this data will be summarized by month, quarter and year.  We will have multiple years of data.  A lot of the reports we will be creating will be in a comparison manner such as 2009 vs 2008 or Jan this year compared to Jan last year or list out sales by month Jan-Dec 2009.  I would like this data to be easily displayed on a report in a side by side manner.  To get this result, what is the best way to structure the data in the tables on a monthly, quarterly and yearly basis?  Right now weu2019ve got one field in the table called date (which is a string) which is listed like:
    Date
    2008YTD
    2009YTD
    2009Jan
    2008Jan
    Is it best to break out the date information so that on the report side it will be easier to work with?  Also should this be set up in the table as a date instead of a string?  If so how do you account for a YTD date?  Are we going to need 2 dates, a start and end date to achieve ytd or qtd information?  Do you recommend creating just a date table and if so how would that be structured?
    So for reporting purposes, using crystal reports, I would like to display comparison data on a report side by side, for this example this year goals compared to last years goals by goal code A-Z (which is a credit code, goals are for the # of credits by code for the year).  The end result I would like is to look like this:
    code   2009 goal   2008 goal
    A        25              20
    B        50              60
    C        10              15
    However the data looks like this (displaying all of the 2009 data first then the 2008 data, not side by side which is how it is in the table):
    code   2009 goal   2008 goal
    A        25
    B        50
    C        10
    etc to Z
    A                          20
    B                          60
    C                          15
    Right now the data is structured in the table like:
    Code  Goal  Date (this is currently a string in the db)
    A        25     YTD 2009
    B        50     YTD 2009
    etc. A-Z for 2009 then:
    A        20      YTD 2008
    B        60      YTD 2008
    Any thoughts on strucuting a table would be appreciated.  thanks.

    Jennifer,
    Most of the DW examples I've seen use a dimDateTime table in the database. That table has multiple columns related to the specific time... For example, here are the columns that are in the, SQL Server sample database, "AdventureWorkdDW"... "DimTime" table
    COLUMN_NAME             COLUMN_INFO
    TimeKey               (int, not null)
    FullDateAlternateKey     (datetime, null)
    DayNumberOfWeek          (tinyint, null)
    EnglishDayNameOfWeek     (nvarchar(10), null)
    SpanishDayNameOfWeek    (nvarchar(10), null)
    FrenchDayNameOfWeek     (nvarchar(10), null)
    DayNumberOfMonth     (tinyint, null)
    DayNumberOfYear          (smallint, null)
    WeekNumberOfYear     (tinyint, null)
    EnglishMonthName     (nvarchar(10), null)
    SpanishMonthName     (nvarchar(10), null)
    FrenchMonthName          (nvarchar(10), null)
    MonthNumberOfYear     (tinyint, null)
    CalendarQuarter          (tinyint, null)
    CalendarYear          (char(4), null)
    CalendarSemester     (tinyint, null)
    FiscalQuarter          tinyint, null)
    FiscalYear          (char(4), null)
    FiscalSemester          (tinyint, null)
    Then all of the fact table receive their date stamps by linking back to this table, using the TimeKey as a foreign key.
    HTH,
    Jason

  • Best way to handle a email subscription and narrow down the emails.

    We are using SCOM 2012 SP1 and have everything updated to UR5.
    We have 2 groups defined.   Machines with names starting with Corp- and Prod-.     We have a subscription to notify on errors with this group.   We get many different alerts that are not necessarily important with
    this subscription so we want to  narrow down the focus.
    We would like to use these two groups but only send out notifications when certain types of events happen like
    CPU utilization is too high or   logical disk space on C: drive is low (dont care as much about D: or E: etc).
    Is there a way to do this with criteria set to specific rules or monitors?
    Thanks Lance

    1) In the 'Subscription' Page there is a criteria "created by specific rules or monitors', use that.
    Get the rules/monitors that generate disk space and other alerts that you want and select them in the above mentioned criteria.
    2) to identify from email weather its for a CORP or PROD server, go to Channels tab and create two email notification channels one for each corp and prod. from the 'Format' page , just add "CORP:"  and "PROD:" in the beginning of the "E-mail subject:"
    option.
    Hope this helps.
    Thanks, S K Agrawal

  • Best way to handle all erros and get performance(OCI)?

    Hi there,
    Im using[b] Oracle Call Interface to execute batch file process. But I have got a problem.
    I set ExecuteBatch with the same number of commit time, i.e: 100, 1000 or just 10, for thats ok.
    ((OraclePreparedStatement)globalStmt).setExecuteBatch(commit);
    And I handle executeUpdate to catch all SQL Exceptions. I made some proposital files with invalid erros but when I handle "executeUpdate" it�s doesn�t get the corrent error line, and puts out another line that is corret.
    ((OraclePreparedStatement)globalStmt).executeUpdate();
    Check in the code I concluded that its always get the same sequence of commit number like the error line. For example, I have between line 1 - 50 a line error, this line is 31, but I set the commit time for 50, its show me that line errror is 50 wherever 31. But if I put setExecuteBatch with '1' so it can handle corret lines erros, but the system performance bring down. What is the best way to handle all erros and keep the perfronace?
    Sorry for my english, I am not native. Thanks all.

    So by doing this, everything will transfer and look exactly the way I have it on the old machine?
    That is correct, if your old machine is Intel based after using MA the new machine will look just like the old machine. Here is information from Apple on MA, I'd recommend looking it over.
    My recommendation is to answer NO when setting up the new machine when it asks "Are you moving from another Mac?" The reason being let you new machine get set up and run for a couple of hours to ensure it's fine. Then launch MA and follow the prompts, it's very easy and if you use a fast connection like FW it should go smoothly.
    Regards,
    Roger

  • Best way to handle large files in FCE HD and iDVD.

    Hi everyone,
    I have just finished working on a holiday movie that my octagenarian parents took. They presented me with about 100 minutes of raw footage that I have managed to edit down to 64 minutes. They have viewed the final version that I recorded back to tape for them. They now want to know if I can put it onto a DVD for them as well. Problem is the FCE HD file is 13Gb.
    So here is my question.
    What is the best way to handle this problem?
    I have spoken to a friend of mine who is a professional editor. She said reduce the movie duration down to about 15mins because it's probably too long and boring. (rather hurtful really) Anyway that is out of the question as far as my oldies are concerned.
    I have seen info on Toast 8 that mentions a "Fit to DVD" process that purports to "squash" 9Gb of movie to a 4.7Gb disk. I can't find if it will also put 13Gb onto a dual layer 8.5Gb disk.
    Do I have to split the movie into two parts and make two dual layer DVD's? If so I have to ask - How come "Titanic", 3hrs+ fits on one disk??
    Have I asked too many questions?

    Take a deep breath. Relax. All is fine.
    iDVD does not look at the size of your video file, it looks at the length. iDVD can accomodate up to 2 hours of movie
    iDVD gives you different options depending on the length of your movie. Although I won't agree with your friend about reducing the length of your movie to 15 minutes, if you could trim out a few minutes to get it under an hour that setting in iDVD (Best Performance though the new version may have renamed it) gives you the best quality. Still, any iDVD setting will give you good quality even at 64 minutes
    In FCE export as Quicktime Movie NOT any flavour of Quicktime Conversion. Select chapter markers if you have them. If everything is on one system unchecked the Make Movie Self Contained button. Drop the QT file into iDVD

  • Best way to handle audio and video?

    Can someone give me suggestions on best way to handle audio and video clips. 2 issues:
    Audio- I want to avoid problems for visitors who may need to install something to listen, and would like them to be able to remain viewing page while listening. I understand how to set up the Iweb player option, and I understand how to simply link to a mp3 file that relies on the visitors player. I'm thinking I need to "embed a player" (?) which is something I haven't dealt with yet.
    Video- I have a link to a .mov file. It plays it in either a real small box, or if you enlarge it , the image deteriorates. Is there a better way to handle this?
    http://www.imaginerecoveryonline.com/IMAGINERECOVERYONLINE.COM/THE_JOURNEY_1.htm l
    Thanks for any help

    Yes that should work for what you want.

  • (workflow question) - What is the best way to handle audio in a large Premiere project?

    Hey all,
    This might probably be suitable for any version of Premiere, but just in case, I use CS4 (Master Collection)
    I am wrestling in my brain about the best way to handle audio in my project to cut down on the time I am working on it.
    This project I just finished was a 10 minute video for a customer shot on miniDV (HVX-200) cut down from 3 hours of tape.
    I edited my whole project down to what looked good, and then I decided I needed to clean up all the Audio using Soundbooth, So I had to go in clip by clip, using the Edit in SoundBooth --> Render and Replace method on every clip. I couldn't find a way to batch edit any audio in Soundbooth.
    For every clip, I performed similar actions---
    1) both tracks of audio were recorded with 2 different microphones (2 mono tracks), so I needed only audio from 1 track - I used SB to cut and paste the good track over the other track.
    2) amplified the audio
    3) cleaned up the background noise with the noise filter
    I am sure there has to be a better workflow option than what I just did (going clip by clip), Can someone give me some advice on how best to handle audio in a situation like this?
    Should I have just rendered out new audio for the whole tape I was using, and then edit from that?
    Should I have rendered out the audio after I edited the clips into one long track and performed the actions I needed on it? or something entirely different? It was a very slow, tedious process.
    Thanks,
    Aza

    Hi, Aza.
    Given that my background is audio and I'm just coming into the brave new world of visual bits and bytes, I would second Hunt's recommendation regarding exporting the entire video's audio as one wav file, working on it, and then reimporting. I do this as one of the last stages, when I know I have the editing done, with an ear towards consistency from beginning to end.
    One of the benefits of this approach is that you can manage all audio in the same context. For example, if you want to normalize, compress or limit your audio, doing it a clip at a time will make it difficult for you to match levels consistently or find a compression setting that works smoothly across the board. It's likely that there will instead be subtle or obvious differences between each clip you worked on.
    When all your audio is in one file you can, for instance, look at the entire wave form, see that limiting to -6 db would trim off most of the unnecessary peaks, triim it down, and then normalize it all. You may still have to do some tweaking here and there, but it gets you much farther down the road, much more easily.Same goes for reverb, EQ or other effects where you want the same feel throughout the entire video.
    Hope this helps,
    Chris

  • Best way to organise multiple large clips and hundreds of subclips

    Situation
    I have the task of producing a suite of short 2 minute videos comprising highlights of several hours of footage in multiple large files.
    I want to extract and manage the highlights in a non-destructive way and organise them according to subject matter.
    I would then use these in several projects to create the highlights videos required.
    I am using Adobe CC, so it's CS6.
    My approach so far is to:
    1. Import all the footage into a single project as individual large clips
    2. Scrub through each clip, set in/outs and extract to subclip, store in subject matter bin
    My questions:
    1.  if I go back to the original footage and edit it, it doesn't seem to flow through to the subclips. How do I make the subclips inherit the attributes of their parent? (e.g. could be sound, colour, etc)
    2. how do I access these highlights bins from another project?
    3. is this the best way to handle my situation?
    Thanks for your input.
    Ric

    I suppose applying the effects to the source clip and rendering out a new source file is one way of doing it.  Just be very careful that you don't compress the output.  And, of course, it will double the amount of hard drive space.  I'd still make all the sub-clips first and use the "paste attributes" command.  I think you can even lasso a whole bunch of clips and paste attributes to all of them at once (they have to be in the timeline, though - not the project panel).
    It shouldn't take an awful lot of hard drive space to duplicate projects.  Just duplicate the "prproj" file, not the whole projects folder (and certainly not the source material).
    Here's a screen shot of my set-up for a complicated, 5-camera shoot of a play:
    The original project (syncing all the cameras up) is 1 MB, each successive archive gets a little bigger but the current project is only 3.8 MB.  I would imagine if you are deleting bins and sequences for the mini-projects, the prproj file would actually get smaller than the original master.
    Okay.. so here's the mantra about non-tape based video... First you back up the card (including the complete folder structure) onto an archive drive.  Next you copy this folder onto an external or RAID or other "real" archive drive.  Then (and only then) do you start editing the material!
    Ideally, you need at least three, physically separate harddrives (not partitions) in an editing station (this is true Windows or Mac).  One drive for the operating system and applications only (no media or projects or anything!).  One for working projects (each project in its own folder) where everything except the source video is stored.  And finally one for the source video.  On my system my OS drive is Tardis ('cuz I'm a geek!).  I have a Projects drive for projects (duh!) a Scratch drive for working video files and an Archive for "finished" projects waiting delivery and back-back-ups of my SD cards.

  • Best way to handle multiple currencies

    I have a requirement that users should be able to report against an OLAP cube in a currency of their choice (from a list of about 20) and was wondering what the best way to handle this might be.
    One option would be to have a currency dimension containing the list of valid currencies and then to pre-calculate measures in each of the currencies and store them in the cube. However the downside of this is that the resultant cube would be 20 times larger than a cube in a single currency, take longer to maintain etc. I could of course partition the cube by currency to improve reporting performance since users would only report in one currency at a time.
    Another alternative would be to dynamically calculate the measures based on exchange rates - I guess this could either be done in the cube itself or as part of the reporting code. However since exchange rates are daily, this would obvioulsy prevent me from aggregating data up the time dimension (all measures are at the day level).
    Is there any standard way of doing this and what are the pro's and con's?
    Thanks,
    Chris

    Sorry - messed up - I should have posted this in the OLAP forum.....

  • Best way to handle multi-page documents?

    Hi,
    I'm creating a multi-page editor on top of TLF and I'd like to hear some opinions about the best way to handle multi-pade documents.
    I'm basicaly considering two options:
    1.) A single RichEditableText and several ContainerControllers.
    2.) Several RichEditableText.
    Tks.

    I'm not sure you can create a RichEditableText and give it multiple ContainerControllers. If you use multiple RichEditableText
    components, then you will have to decide in advance which text goes in which one.
    We have posted sample code for doing this in straight ActionScript, which you could presumably also host inside a Flex container. See:
    http://blogs.adobe.com/tlf/examples/
    Look for "ActionScript Pagination Example".
    - robin

  • OIM 9031: Best way to handle application/test account on target system.

    Hi Guys,
    I am wonder what will be the best way to handle application account created in target sytem . i.e. I have target system Active Directory and on non-trusted reconciliation, I also fetch application/test account which not going to match to any existing user ,but should be capture for reporting or any future action .
    Any input or idea is most welcome !!
    Cheers,
    Ankit

    There are basically two approaches to handle service accounts.
    Either you model them as a free standing RO very similar to a normal AD account or you use the built in "service account" and associate the account with an already existing AD RO instance. I haven't used the "service account" approach in any customer project yet so I can't really comment on the details of that approach (hopefully someone else will be able to do that).
    Are you sure that you have service accounts in AD that you can't attribute to a specific users? Most organizations require service accounts to be linked to a user or a group of users so that the need for the account to continue existing can be verified by a human. Having live accounts in your AD that no one can say what they do or why they exist is normally a very scary thought for most organizations.
    Hope this helps
    /M

  • What is the best way to handle very large images in Captivate?

    I am just not sure the best way to handle very large electrical drawings.
    Any suggestions?
    Thanks
    Tricia

    Is converting the colorspace asking for trouble?  Very possibly!  If you were to do that to a PDF that was going to be used in the print industry, they'd shoot you!  On the other hand, if the PDF was going online or on a mobile device – they might not care.   And if the PDF complies with one of the ISO subset standards, such as PDF/X or PDF/A, then you have other rules in play.  In general, such things are a user preference/setting/choice.
    On the larger question – there are MANY MANY ways to approach PDF optimization.  Compression of image data is just one of them.   And then within that single category, as you can see, there are various approaches to the problem.  If you extend your investigation to other tools such as PDF Enhancer, you'd see even other ways to do this as well.
    As with the first comment, there is no "always right" answer.  It's entirely dependent on the user's use case for the PDF, requirements of additional standard, and the user's needs.

  • Best way to remove last line-feed in text file

    What is the best way to remove last line-feed in text file? (so that the last line of text is the last line, not a line-feed). The best I can come up with is: echo -n "$(cat file.txt)" > newfile.txt
    (as echo -n will remove all trailing newline characters)

    What is the best way to remove last line-feed in text file? (so that the last line of text is the last line, not a line-feed). The best I can come up with is: echo -n "$(cat file.txt)" > newfile.txt
    (as echo -n will remove all trailing newline characters)
    According to my experiments, you have removed all line terminators from the file, and replaced those between lines with a space.
    That is to say, you have turned a multi-line file into one long line with no line terminator.
    If that is what you want, and your files are not very big, then your echo statement might be all you need.
    If you need to deal with larger files, you could try using the 'tr' command, and something like
    tr '
    ' ' ' <file.txt >newfile.txt
    The only problem with this is, it will most likely give you a trailing space, as the last newline is going to be converted to a space. If that is not acceptable, then something else will have to be arranged.
    However, if you really want to maintain a multi-line file, but remove just the very last line terminator, that gets a bit more complicated. This might work for you:
    perl -ne '
    chomp;
    print "
    " if $n++ != 0;
    print;
    ' file.txt >newfile.txt
    You can use cat -e to see which lines have newlines, and you should see that the last line does not have a newline, but all the others still do.
    I guess if you really did mean to remove all newline characters and replace them with a space, except for the last line, then a modification of the above perl script would do that:
    perl -ne '
    chomp;
    print " " if $n++ != 0;
    print;
    ' file.txt >newfile.txt
    Am I even close to understanding what you are asking for?

  • Best way to handle text files in OD10g

    We have a requirement to store reports in text format into a database field, to be able to view the reports, and to print them if desired using Forms 10g. What is the best way to handle this?
    - define the field in the database as clob or blob?
    - if CLOB is the choice, what tools to use to upload CLOBs to the database (since webutil transfer is for blob only)?
    - in Forms 10g, can one use the Forms data type LONG for CLOB?
    - can you do Forms search on clob and blob fields?
    - how can reports that are stored in fields be viewed without first downloading to the client workstation?
    - in Forms 10g, what is the best way to view text files residing in local PCs: "host notepad myFile"?
    Thanks much for your reply!
    gk

    Take a deep breath. Relax. All is fine.
    iDVD does not look at the size of your video file, it looks at the length. iDVD can accomodate up to 2 hours of movie
    iDVD gives you different options depending on the length of your movie. Although I won't agree with your friend about reducing the length of your movie to 15 minutes, if you could trim out a few minutes to get it under an hour that setting in iDVD (Best Performance though the new version may have renamed it) gives you the best quality. Still, any iDVD setting will give you good quality even at 64 minutes
    In FCE export as Quicktime Movie NOT any flavour of Quicktime Conversion. Select chapter markers if you have them. If everything is on one system unchecked the Make Movie Self Contained button. Drop the QT file into iDVD

Maybe you are looking for

  • Adobe LiveCycle Designer 9.8

    Hello, I would like to install the latest version of Adobe LiveCycle Designer. According to SAP Note 1658830 ( 'Adobe LiveCycle Designer 9.8 - Installation' ) this is the version 9.8. I have searched the file "67837800102000002446 Adobe LiveCycle Des

  • Protected Flex field and Custom Views

    Hi, I have created some project variables in oracle bpm 11.1.1.6.0 project.After deployment, these variables are coming as Protected flex field but for some reason these are not visible when I define custom views. I also tried creating Protected flex

  • Error in trigger (simple)

    hi , I 've written the following trigger but unable to trace why the following errors .. All the columns are written properly as per the table ... CREATE OR REPLACE TRIGGER TRI1 AFTER INSERT OR UPDATE OR DELETE OF INL_ID,MRR_CD, REQ, MM, ML_STS, FLL_

  • Workflow not trigered in the MDM:POST Process Material.

    Hi Workflow Experts,             I have copied the standard workflow ws47300003.This workflow is trigeered whenever the material has been changed and send mail to respective user.            Here i have  a problem that, after changing some data in ma

  • ASE 15.7 Job Scheduler won´t start again

    Hi, we encoutered the following problem in our ECC6.0 / EHP5 on ASE 15.7 PL 122 System: DBACockpit / Collector Configuration shows the warning "The ASE-Job Scheduler is not active". The log SID_JSAGENT shows the following entries: 00:11704:12104:2014