Init Data Upload from ODS to Info cube

Hi,
When i am trying to upload the date from ODS to info cube using initial upload option in info package, i am not able to see the selection option fields.
Whereas using the same ODS & info cube, if i am trying to change the option to full upload in info package, i am getting all the selection fields.
what could be the reason for the same.
please help
Rajiv

Somnath,
Let me explain u the complete scenario
1) we have an existing ODS which populates an info cube on daily basis (delta upload)
2) we have created a new info cube, which fetch the data from the same ODS. that means 2 info cubes which are getting populated from one ODS.
3) now when we have tried to initialize the new info cube from the existing ODS, it asked us to delete the initialize upload options "scheduler - initialize option for the system". & we have done the same. but by doing this, ODS has removed the option of delta upload from old info cube upload info package &&  new info cube initial upload is not dislaying any of selection fields..
4) now, i have a query
- how can i activate delta on old info cube (without losing any data)
- how to upload data in new info cube with selection fields and start delta
Regards
Rajiv

Similar Messages

  • Start routine from ODS to Info cube in bw 3.5

    Hi Folks,
                     We have a code for START ROUTINE from ODS to Info Cube.
    we are doing FULL Load to ODS(0CRM_QUT1) and from ODS to INFO CUBE also we are doing FULL Load.
    Iam planning to change to DELTA Load from ODS to Info Cube to improve performance.
    Is there any situtation If  we write a Start routine from ODS to Info cube we can't do DELTA Load .
    Please clarify me.

    improving performance is a good thing. using delta load mechanism is also good for performance as loadings are faster and are less time cosuming. Nevertheless analyse carefully how the solution was built to understand clearly why a full load was required at the time of the design of the solution. maybe the full load was used for some reasons that are still valid or would require more than just changing from a full to a delta load . in conclusion it is always good to use delta but in your case look carefully at the existing coding it is possible that look up is coded for inserting records or other treatments.
    hope this could help you out.

  • Data loads from ODS TO CUBE

    Hi,
       i have delta loads comming into the ODS. i do full update from ODS to the cube by date range for material moments no. last time when i loaded the data, it loaded few for the date range. rest did not load and sitting at ODS. this is full load and tried to load again. any suggestions...
    sp

    Hi Srinivas,
            check your update rules between ODS and cube whether they are mapped properly(check your date range for the cube load).
             Do a Init load and then do the delta load.
    hope this will help.

  • Summing up key figure in Cube - data load from ODS

    Hello Gurus,
    I am doing a data load from ODS to cube.
    The records in ODS are at line-item level, and all needs to be summed up at header level. There is only one key-figure. ODS has header and line-item fields.
    I am loading only header field data ( and not the item-field data) from ODS to Cube.
    I am expecting only-one record in cube ( for all the item-level records) with all the key-figures summed up. But couldn't see that.
    Can anyone please explain how to achieve it.
    I promise to reward points.
    =====
    Example to elaborate my point.
    In ODS
    Header-field  item-field   quantity
    123                301          10
    123                302           20
    123                303           30
    Expected record in Cube
    Header-field       Quantity
       123                    60  
    ====================
    Regards,
    Pramod.

    Hello Oscar and Paolo.
    Thanks for the reply. I am using BW 7.0
    Paolo suggested:
    >>If you don't add item number to cube and put quantity as adition in update rules >>from ODS to cube it works.
    I did that still I get 3 records in cube.
    Oscar Suggested:
    >>What kind of aggregate do you have for your key figure in update rules (update >>or no change)?
    It is "summation". And it cannot be changed. (Or at least I do not know how to change it.)
    There are other dimensions in the cube - which corresponds to the field(s) in ODS.
    But, I just mentioned these two (i.e. header and item-level) for simplicity.
    Can you please help?
    Thank you.
    Pramod.

  • Delta records are not loading from DSO to info cube

    My query is about delta loading from DSO to info cube. (Filter used in selection)
    Delta records are not loading from DSO to Info cube. I have tried all options available in DTP but no luck.
    Selected "Change log" and "Get one request only" and run the DTP, but 0 records got updated in info cube
    Selected "Change log" and "Get all new data request by request", but again 0 records got updated
    Selected "Change log" and "Only get the delta once", in that case all delta records loaded to info cube as it was in DSO and  gave error message "Lock Table Overflow" .
    When I run full load using same filter, data is loading from DSO to info cube.
    Can anyone please help me on this to get delta records from DSO to info cube?
    Thanks,
    Shamma

    Data is loading in case of full load with the same filter, so I don't think filter is an issue.
    When I follow below sequence, I get lock table overflow error;
    1. Full load with active table with or without archive
    2. Then with the same setting if I run init, the final status remains yellow and when I change the status to green manually, it gives lock table overflow error.
    When I chnage the settings of DTP to init run;
    1. Select change log and get only one request, and run the init, It is successfully completed with green status
    2. But when I run the same DTP for delta records, it does not load any data.
    Please help me to resolve this issue.

  • Open Orders are negitive in ODS and Info cube

    Hi,
    Our ODS is getting data from sales order Item data and from ODS data is going to Info cube.
    The problem is that: for few sales orders some open orders are deleting in R/3 but same records are reversing
    the old records due to this in the ODS and info cube sales orders Quantity is negative but the orders are not negitive.
    Could any one let me know how do we can ressolve it.
    Regards,
    Sharma. IVN

    Hi Sarma,
    You should consider checking the attached links below:
    ROCANCEL field in R3 extraction program can't catch the LOEKZ (deletion ind
    http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/oss_notes/sdn_oss_bw_bct/~form/handler%7b5f4150503d3030323030363832353030303030303031393732265f4556454e543d444953504c4159265f4e4e554d3d383836373136%7d
    This may help in this case.
    Regards,
    Pietro

  • Data upload from R3 to BPC using filters

    We are facing a performance issue on the uploading data flow from R3 to BPC. Two steps have been set up.
    Step1: R/3 --> ODS --> BW CUBE (No filters and full mode)
    Step2: BW  CUBE --> BPC CUBE (Standard BW upload package)
    The first one (from R3 to a BW Cube) has no filters, so it loads every data and it takes too much time. However, the second one (from the Cube to BPC)  has filtering options, so we can only load the data we need to (we usually use entity and time filters).
    Both are executed from the BPC DataManager.
    Any advice to improve the performance of the first step? Could we join the step 1 and step 2 and get the filters from the standard BW upload process using a custom process chain?
    Does anybody have any experience using filters to reduce the amount of records on the data uploading from R/3 to BPC?
    BPC Version 7.50.15

    Hi,
    for example in BI data is stored as below
    GroupAcc - Date - Balance
    100000 - 12.12.10 - 400
    **I'm not quite sure how you are getting the above data from infocube.Are you directly loading data from PSA to infocube with out loading data into DSO -0FIGL_O10(General ledger(New):Transaction figures) ?
    To get required format of following data, no need to do any thing while laoding the data into BPC. Your BI infocube should maitain the data in below format.
    GroupAcc - Date - Balance
    100000 - 12.12.10 - 100 - (LC)
    100000 - 12.12.10 - 100 - (TC)
    100000 - 12.12.10 - 200 - (GC)
    To acheive above format of data in the infocube please follow standard way of data flow given by SAP in BI Contect(http://help.sap.com/saphelp_nw70ehp1/helpdata/en/e6/f16940c3c7bf49e10000000a1550b0/frameset.htm).
    Flow should be
    R/3 --> PSA --> DSO -->Infocube.
    In this flow its important to remember about infoobect 0CURKEY_TC in DSO.
    This is always the currency key of the transaction currency; it is also filled for records with a different currency type. Without this key field, postings with different transaction currencies would be overwritten after summarization and thereby be lost.
    (http://help.sap.com/saphelp_nw70ehp1/helpdata/en/a8/e26840b151181ce10000000a1550b0/content.htm)
    Once you follow the standard flow, your infocube contains the requried format (LC,TC and GC )of data to load it into BPC application.
    hope it helps...
    regards,
    Raju

  • Data loading from flat file to cube using bw3.5

    Hi Experts,
                       Kindly give  me the detailed steps with screens  about Data loading from flat file to cube using bw3.5
           ...............Please

    Hi ,
    Procedure
    You are in the Data Warehousing Workbench in the DataSource tree.
           1.      Select the application components in which you want to create the DataSource and choose Create DataSource.
           2.      On the next screen, enter a technical name for the DataSource, select the type of DataSource and choose Copy.
    The DataSource maintenance screen appears.
           3.      Go to the General tab page.
                                a.      Enter descriptions for the DataSource (short, medium, long).
                                b.      As required, specify whether the DataSource builds an initial non-cumulative and can return duplicate data records within a request.
                                c.      Specify whether you want to generate the PSA for the DataSource in the character format. If the PSA is not typed it is not generated in a typed structure but is generated with character-like fields of type CHAR only.
    Use this option if conversion during loading causes problems, for example, because there is no appropriate conversion routine, or if the source cannot guarantee that data is loaded with the correct data type.
    In this case, after you have activated the DataSource you can load data into the PSA and correct it there.
           4.      Go to the Extraction tab page.
                                a.      Define the delta process for the DataSource.
                                b.      Specify whether you want the DataSource to support direct access to data.
                                c.      Real-time data acquisition is not supported for data transfer from files.
                                d.      Select the adapter for the data transfer. You can load text files or binary files from your local work station or from the application server.
    Text-type files only contain characters that can be displayed and read as text. CSV and ASCII files are examples of text files. For CSV files you have to specify a character that separates the individual field values. In BI, you have to specify this separator character and an escape character which specifies this character as a component of the value if required. After specifying these characters, you have to use them in the file. ASCII files contain data in a specified length. The defined field length in the file must be the same as the assigned field in BI.
    Binary files contain data in the form of Bytes. A file of this type can contain any type of Byte value, including Bytes that cannot be displayed or read as text. In this case, the field values in the file have to be the same as the internal format of the assigned field in BI.
    Choose Properties if you want to display the general adapter properties.
                                e.      Select the path to the file that you want to load or enter the name of the file directly, for example C:/Daten/US/Kosten97.csv.
    You can also create a routine that determines the name of your file. If you do not create a routine to determine the name of the file, the system reads the file name directly from the File Name field.
                                  f.      Depending on the adapter and the file to be loaded, make further settings.
    ■       For binary files:
    Specify the character record settings for the data that you want to transfer.
    ■       Text-type files:
    Specify how many rows in your file are header rows and can therefore be ignored when the data is transferred.
    Specify the character record settings for the data that you want to transfer.
    For ASCII files:
    If you are loading data from an ASCII file, the data is requested with a fixed data record length.
    For CSV files:
    If you are loading data from an Excel CSV file, specify the data separator and the escape character.
    Specify the separator that your file uses to divide the fields in the Data Separator field.
    If the data separator character is a part of the value, the file indicates this by enclosing the value in particular start and end characters. Enter these start and end characters in the Escape Charactersfield.
    You chose the; character as the data separator. However, your file contains the value 12;45 for a field. If you set u201C as the escape character, the value in the file must be u201C12;45u201D so that 12;45 is loaded into BI. The complete value that you want to transfer has to be enclosed by the escape characters.
    If the escape characters do not enclose the value but are used within the value, the system interprets the escape characters as a normal part of the value. If you have specified u201C as the escape character, the value 12u201D45 is transferred as 12u201D45 and 12u201D45u201D is transferred as 12u201D45u201D.
    In a text editor (for example, Notepad) check the data separator and the escape character currently being used in the file. These depend on the country version of the file you used.
    Note that if you do not specify an escape character, the space character is interpreted as the escape character. We recommend that you use a different character as the escape character.
    If you select the Hex indicator, you can specify the data separator and the escape character in hexadecimal format. When you enter a character for the data separator and the escape character, these are displayed as hexadecimal code after the entries have been checked. A two character entry for a data separator or an escape sign is always interpreted as a hexadecimal entry.
                                g.      Make the settings for the number format (thousand separator and character used to represent a decimal point), as required.
                                h.      Make the settings for currency conversion, as required.
                                  i.      Make any further settings that are dependent on your selection, as required.
           5.      Go to the Proposal tab page.
    This tab page is only relevant for CSV files. For files in different formats, define the field list on the Fields tab page.
    Here you create a proposal for the field list of the DataSource based on the sample data from your CSV file.
                                a.      Specify the number of data records that you want to load and choose Upload Sample Data.
    The data is displayed in the upper area of the tab page in the format of your file.
    The system displays the proposal for the field list in the lower area of the tab page.
                                b.      In the table of proposed fields, use Copy to Field List to select the fields you want to copy to the field list of the DataSource. All fields are selected by default.
           6.      Go to the Fields tab page.
    Here you edit the fields that you transferred to the field list of the DataSource from the Proposal tab page. If you did not transfer the field list from a proposal, you can define the fields of the DataSource here.
                                a.      To define a field, choose Insert Row and specify a field name.
                                b.      Under Transfer, specify the decision-relevant DataSource fields that you want to be available for extraction and transferred to BI.
                                c.      Instead of generating a proposal for the field list, you can enter InfoObjects to define the fields of the DataSource. Under Template InfoObject, specify InfoObjects for the fields in BI. This allows you to transfer the technical properties of the InfoObjects into the DataSource field.
    Entering InfoObjects here does not equate to assigning them to DataSource fields. Assignments are made in the transformation. When you define the transformation, the system proposes the InfoObjects you entered here as InfoObjects that you might want to assign to a field.
                                d.      Change the data type of the field if required.
                                e.      Specify the key fields of the DataSource.
    These fields are generated as a secondary index in the PSA. This is important in ensuring good performance for data transfer process selections, in particular with semantic grouping.
                                  f.      Specify whether lowercase is supported.
                                g.      Specify whether the source provides the data in the internal or external format.
                                h.      If you choose the external format, ensure that the output length of the field (external length) is correct. Change the entries, as required.
                                  i.      If required, specify a conversion routine that converts data from an external format into an internal format.
                                  j.      Select the fields that you want to be able to set selection criteria for when scheduling a data request using an InfoPackage. Data for this type of field is transferred in accordance with the selection criteria specified in the InfoPackage.
                                k.      Choose the selection options (such as EQ, BT) that you want to be available for selection in the InfoPackage.
                                  l.      Under Field Type, specify whether the data to be selected is language-dependent or time-dependent, as required.
           7.      Check, save and activate the DataSource.
           8.      Go to the Preview tab page.
    If you select Read Preview Data, the number of data records you specified in your field selection is displayed in a preview.
    This function allows you to check whether the data formats and data are correct.
    For More Info:  http://help.sap.com/saphelp_nw70/helpdata/EN/43/01ed2fe3811a77e10000000a422035/content.htm

  • Problem when loading from ODS to the CUBE

    Hi Experts,
    I am facing an unusual problem when loading data from ODS to the cube.
    I have a first level ODS where the delta postings are updated. I have checked the active table of the ODS and the data is accurate.
    I have deleted the entire data from the Cube and trying to do a full load from ODS to the CUBE.
    I am sure when I run a full load the data goes from Active table of the ODS.
    After the full load the the keyfigure values are 0. I have tried doing the testing by loading couple of sales documents and still the keyfigure values are 0.
    I wonder when I load the full. The data should be picked exactly the way it is in active table of the ODS.
    I also dont have any fancy routines in the update rules.
    Please help me in this regard.
    Regards
    Raghu

    Hi,
    Check the procedure did u do that exactly or not. just follow the the laymen steps here:... and let me know ur issue
    o     First prepare a flat file in Microsoft excel sheet for master data and transaction data and save it in .CSV format and close it.
    o     First select info objects option and create info area then create info object catalog then char, & key figure. Then create ‘id’ in Char, Name as attribute activate it then in key figures create no and activate it.
    o     In info sources create application component then create info source with direct update for master data and flexible update for transaction data. The flat files would be created for master data create info package and execute it.
    o     For transaction data go to info provider then select the file right click on it then select option create ODS object. Then a new screen opens give the name of the ODS and create other screen opens then drag character to key fields.
    o     Activate it then update rules it will give error to it then go to communication structure and give 0Record mode then activate it then go to update rules it will activate it.
    o     Then go to info source then create info package then external data then processing update, Data targets then the scheduling then click start then the monitor.
    o     Then other screen opens there we can see if we doesn’t able to see the records then come back to the info provider right click on it then other screen opens then select the option activate data in ODS then a screen opens.
    o     so select the option QM where it would be in a yellow co lour then by double click then another screen opens then select green color option then click continue it and it comes to the previous screen then select the option and click save it would be underneath of it just click it up and close it.
    o     Then come back to the main screen to the info source to the info package go to the package and right click on it then a screen opens then click schedule without distributing any of the tab directly click the monitor tab. Then refresh it until the records come to the green.
    o     Once it is green the click the data target and see the file executed.
    o     Now go to the info provider then a right click on the file name then create info cube. Then a new screen opens and gives the name and activates it. The drag the key figures to the other side, time characteristics to the other side, and chacetristics to the other side.
    o     Then create the dimension table then assign it and activate it .Then go to file name right click on it then select the option create update rules select ODS option and give the file name and activate it .
    o     Then come back to the main screen and refresh it. Whenever you see 8 file names you will know that it is undergoing the data mart concept.
    o     In the main screen click ODS file right click then a pop up screen opens select the option update ODS data target. Then another screen opens in that we are able to see 2 options full update, initial update then select initial update.
    o     Then another screen opens info Package there the external data , data targets , scheduler select the option later in background option then a another screen opens just click immediate then click save automatically it closes then click start option.
    o     Then select monitor option then the contents then the field to be elected then the file to be executed.
    regards
    ashwin

  • Update data targets from ODS object automatically

    Hello experts,
    I have an ods (that has two delta loading every night) with the setting: "Update data targets from ODS object automatically". This ODS is source of only one other ODS.
    But we have a chain when there are the seguent steps: delta 1 --> ODS1,
    delta 2 --> ODS1; Activate ODS1; delta --> ODS2; Activate ODS2.
    The setting "Update data targets...automatically" could create some problems?
    I think this because it is happened that in delta loading of ODS2, some records have only the key filled, and all the data fields blank; then, if I make a full repair of these records, all data are correctly filled.
    Could be this the cause?
    Help me..it is urgent..
    Thanks,
    Vito

    Hi Vito,
    To have better control of processes when ods gets data from multiple sources and supply it forward to ODS and cube ...we generally deactivate the flag "update data automatically".
    And Process chains looks like ..
    From source 1,2,3 update data into ODS 1.
    Activate Data ODS1
    Initiate delta from ODS1 to ODS2.
    Activate data ODS2.
    We did face some problem of blank records with only key filled and later deactivated the checbox "Update data into data targets automatically" and it resovled lots of issues.
    Hope that helps.
    Regards
    Mr Kapadia
    Assigning points is the way to say thanks in SDN.

  • Update data targets from ODS

    Hi Friends,
    I have question regarding how to update data further from ODS.
    I have ODS1 and ODS2.
    ODS1 is having deltas updated from ODSA.
    ODS2 is a new ODS and now needs to be updated from ODSA.
    I have reviewed the help already.
    http://help.sap.com/saphelp_nw04/helpdata/en/2c/749bbb595511d4b2bc0050da4c74dc/content.htm
    I want to initially update the ODS2 with <u>selection conditions</u> from ODSA.
    After that I want to load delta from ODSA to ODS2 by automating the process in process chain.
    Please help me know, What is the optimum solution and what are the steps need to be executed.
    Thanks,
    John.

    Anil,
    I am just repeating the procedure to cinfirm that I understand it properly.
    1. Go to scheduler, Initialization option make the status of the overall request red.
    2. when it asks for update the QM status to all data targets, select no.
    3.and then delete the init. request from Scheduler.
    By doing this we just deleting the initialization selection and the request in the subsequent data targets are still green and they contain data.
    After that Init. without data transfer will do the work.
    Please confirm,
    I really appreciate your input on this.
    Thanks.

  • The load from ODS - 0FIAP_O03 to cube 0FIAP_C03  is duplicating records

    Hello Everyone
    I need some help/advice for the following issue
    SAP BW version 3.5
    The Delta load for ODS - 0FIAP_O03 works correctly
    The load from ODS - 0FIAP_O03 to cube 0FIAP_C03  is duplicating records
    NB Noticed one other forum user who has raised the same issue but question is not answered
    My questions are 
    1. Is this a known problem?
    2. Is there a fix available from SAP?
    3. If anyone has had this issue and fixed it, could you please share how you achieved this
    i have possible solutions but need to know if there is a standard solution
    Thankyou
    Pushpa

    Hello Pushpa,
    I assume that you are using the Delta load to the initial ODS and then to the CUBE as well.
    If the Delta is placed in both the places then there should not be any issue while sending the data to CUBE.
    If you are using the FULL Load then normally the data will gets aggregated in the cube as objects are with the Addition mode in the Cube.
    Can you post the exact error that you are facing here as this also can be of the design issue.
    Murali

  • How to schedule Job for data uploading from source to BI

    Hi to all,
    How to schedule Job for data uploading from source to BI,
    Why we required and how we do it.
    As I am fresher in BI, I need to know from bottom.
    Regards
    Pavneet Rana

    Hi.
    You can create [process chain |http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/502b2998-1017-2d10-1c8a-a57a35d52bc8?quicklink=index&overridelayout=true]for data loading pocess and schedule start process to any time/date etc ...
    Regadrs.

  • Error on activation of data records from ODS object

    hi,
    Kindly help with my BW Problem, the error occured on the
    activation of ODS. The error says "<i>Request
    REQU_43MNOPW29W4F5M037J7OAFP52 in ODS ZPUR_O01 must have QM
    status green before it is activated</i>" and "<i>Activation of data
    records from ODS object ZPUR_O01 terminated</i>"
    I tried to check the data in the ODS Object by choosing
    <LS>Administrator Workbench > data targets > context menu on ODS
    Object ZPUR_O01 but the status are green and when i tried to
    activate it the logs says "<i>ODS object ZPUR_O01 was built
    incorrectly. Cannot update request
    REQU_43MODQ3GF69X17B71ZKQMN85I(20,154)
    ODS object ZPUR_O01 was built incorrectly. Cannot update request
    REQU_43MQ8A8F0L1US5COQTUUWX3BA(20,155)
    ODS object ZPUR_O01 was built incorrectly. Cannot update request
    REQU_43MQ8OQQTY21YQ37RMA1WMO06(20,156)
    Data to be activated successfully checked against archiving
    objects                                   
    Activation of data records from ODS object ZPUR_O01 terminated</i>"
    Please help me with this problem.

    Hi Jay
    U need to update these requests to ODS
    REQU_43MODQ3GF69X17B71ZKQMN85I(20,154)
    REQU_43MQ8A8F0L1US5COQTUUWX3BA(20,155)
    REQU_43MQ8OQQTY21YQ37RMA1WMO06(20,156)
    1) delete the current request
    2)Reconstruct the requests above three request
    3)and then Load the current reqest (REQU_43MNOPW29W4F5M037J7OAFP52)
    Then only u will be able to activate the ODS....
    This is an serialization issue...
    hope it helps
    regrds
    AK
    assign points if usefull

  • Error:Activation of data records from ODS object terminated

    Hi All
    i got error while loading the data "Activation of data records from ODS object terminated '
    could you please tell me procedure how can i solve this issue?
    regards
    vasu.

    - First check the Error Messages from the context menu of the activation step. Also check ST22 & SM21. This will tell you where the error is. Sometimes you might get duplicate records.
    - Next use the "Manage" context menu for the ODS in RSA1. Check if any of the previous records are red. If its red, it would mean that either load/activation has failed for previous request which needs to be corrected first b4 you correct the current one.
    - If al the above are OK, simply repeat the Activation step.
    Hope this helps.
    Regards
    Anujit Ghosh

Maybe you are looking for

  • Is there away to find out which account an app assoicated with in itunes

    iTunes is telling two things: - In the left hand colum Apps indicates one [1] app has an update. - In the right hand window it says "No updates are currently available. To check for updates for another Apple ID, sign in with that Apple ID I have logg

  • Pdf create problem

    i generate the pdf with tamil "latha.ttf" font for some values using FOP i connot get clear result anybody know please send related topic guide thanks with regards iyyadurai

  • What is the best device to transfer VHS to iMovie for edit

    Currently have a mac pro and would like to edit vhs videos of the family, did try ezecap waste of money only worked on pc and no sound would like to copy direct to imovies

  • Iphone not charging from jaguar ipod dock in car

    i have a jaguar x-type 2003 with the audio connectivity module added on. It has a dock connector. When I attach my iPhone, it prompts airplaneode and says it is an unsupported device. If I click no to airplane mode, it works through the stereo to lis

  • Any options for a non-mac-compatible printer?

    My Samsung ML-1710 printer is not Mac-compatible (says so clearly on their website). No other options? Just throw my printer out the window and get a new one?