I can't build an xsd for a flat file (txt) to handle repeating records

Hi - have looked at many posts around flat file schema and they don't seem to address my question.
I have a flat file that is \n delimited
the pattern of the data is simple:
record1 - 90 characters
record2 - 20 characters
record3 - n 248 characters - each of these records is parsed into children by the positional method
record n+1 10 characters
record n+2 20 characters
so I used the flat file schema generator to generate the schema and built a map mapping the flat file schema to another xml schema. The schema looks ok - record1, record2, record n+1, record n+2 are child elements of the root. the repeating record
section is showing up as a node with the parsed children.
The transform is only mapping the children of the repeating records. When I test the map only the first repeating record gets parsed. No repeating happens (the actual flat file has 400+ repeating records). When I run the map in debug mode, the input
xml shows that record1 is read in correctly, record2 is read in correctly, record3 is read in and parsed and record4 is treated like record n+1 and record5 is treated like record n+2 and the map thinks it's all finished.
the section of the repeat part of the schema is and you can see that I set the minOccurs=1 and maxOccurs=unbounded for the node (INVOICE) and the complexType but this is not an affective syntax. I have looked at how the EDI X12 schema look and how they handle
looping and it is a lot different than what the Flat File schema wizard is doing. Is there a good set of rules published that would guide me though this? otherwise I will basically have to read in the lines from the file and parse them out with functoids -
seems so inelegant. Thanks in advance.
<xs:element minOccurs="1" maxOccurs="unbounded" name="INVOICE">
          <xs:annotation>
            <xs:appinfo>
              <b:recordInfo structure="positional" sequence_number="3" preserve_delimiter_for_empty_data="true" suppress_trailing_delimiters="false"
/>
            </xs:appinfo>
          </xs:annotation>
          <xs:complexType>
            <xs:sequence minOccurs="1" maxOccurs="unbounded">
              <xs:annotation>
                <xs:appinfo>
                  <groupInfo sequence_number="0" xmlns="http://schemas.microsoft.com/BizTalk/2003"
/>
                </xs:appinfo>
              </xs:annotation>
              <xs:element name="SegmentType" type="xs:string">
                <xs:annotation>
                  <xs:appinfo>
                    <b:fieldInfo justification="left" pos_offset="0" pos_length="2" sequence_number="1" />
                  </xs:appinfo>
                </xs:annotation>
              </xs:element>....... more children elements
Harold Rosenkrans

Thanks for responding
I gave up trying to parse the repeating record into fields. Instead I just loop through the repeating record section with an <xs:for-each> block in the xsl and use functoids to grab the fields.
So that works for having the two, shorter header records (structure is positional) before the section of repeating records. Now I just have to figure out how to get the schema to handle the two, shorter trailer (or footer, whatever you prefer) records after
the section of repeating records
the error I get in VS when I test the map is [BTW I changed the element names in the schema which is why you don't see INVOICE in the error]
When I declare the last element as being positional with a character length of 10 I get the error:
Error 18 Native Parsing Error: Unexpected end of stream while looking for:
'\r\n'
The current definition being parsed is SAPARData. The stream offset where the error occured is 1359. The line number where the error occured is 9. The column where the error occured is 0. 
so the first record is 77 char in length and the second is 16 char and then the repeating records (5 in the file) are 248 char and the last record is 10 char
so an offset of 1359 puts it beyond the last record by 16 characters - so the stream reader is looking for the next repeating record.
if I try to declare the last element as delimited I get the error:
Error 14 Native Parsing Error: Unexpected data found while looking for:
'\r\n'
The current definition being parsed is SAPARData. The stream offset where the error occured is 597. The line number where the error occured is 5. The column where the error occured is 0. 
so the first record is 77 char in length and the second is 16 char and then the repeating records are 248 char.
a stream offset of 597 puts me 8 characters into the third repeating record - at this point I have only declared one trailer record in the  schema, 10 characters long.
Why is stream reader stopping at such a weird spot?
The bottom line is I still haven't discovered the correct schema to handle the trailer records. even if I set the maxOccurs="4" (for the repeat record declaration) it still gets the first error. How does it find an unexpected end of stream looking
for \r\n when the maxOccurs for the repeat record declaration should have the stream pointer in the 5th repeat record.
I unfortunately don't have any options concerning the file structure.
I have read a lot of posts concerning the trailer issue. I have seen a couple that looked interesting. I guess I'll just have to give them a try. The other option is to create a custom pipeline that will only take file lines of 248 characters.
That's just disgusting !
Harold Rosenkrans

Similar Messages

  • Can you build a battery for iphone that will last till the end of the day?

    Can you build a battery for iphone that will last till the end of the day?

    No, I can't. I've never taken any classes in either electrical or chemical engineering.
    While you're looking for someone who can, you may want to try searching the forums for the many other threads that give tips on extending your battery life (things like turning off Background Refresh). You may also want to look at battery cases such as those made by Mophie.
    Best of luck.

  • XSD for the input file for GTS offline screening program

    Hi Friends,
    We are running some performance tests for our current project to  evaluate how much time the offline screening program (report "/SAPSLL/SPL_OFFLINE_SCREENING") takes in GTS. This report expects the customers in an XML format - would you know where I can get the definition of this expected XML format (e.g. the XSD for this input file)?
    Can some one help me to provide response XML definition of what GTS generates for the SPL check ?
    Thanks,
    Sudheer.

    Hi Sudheer,
    please follwow the steps:
    1. Prepare an XML file  with Adress data
    -> you can create a simple excel file that has to be converted into an      
    XML-file. See also the GTS documentation http://help.sap.                
    com/saphelp_gts72/helpdata/en/b1/8dc5c62c885046ab71fdcddc6ad2c2/frameset 
    .htm.                                                                    
    2. Upload the local file  to GTS for SPL Screening
    3. Examine the SPL screening results in GTS
    4. Save the results in Audit trail and&or in local file
    Best regards,
    Christin

  • Can we handle content convertion for TabDelimiter flat file in JMS

    Hi,
    Is it possible to handle content conversion for TabDelimiter flat file in JMS (Not in FILE adapter).
    If possible .. can you please explain.
    Regards,
    Siva.

    Hi, Siva:
    Yes, this is possible.
    The way that JMS adapter to handle content cnversion is similar to the way file adapter does.
    Different from file adapter, you do not have to specify "Content Conversion" as transport protocol, you need to use transformation bean (localejbs/AF_Modules/MessageTransformBean) in Module tab.
    Check page 23 of the following link:
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/a05b2347-01e7-2910-ceac-c45577e574e0
    Liang

  • Can we cleanse and transform data at flat file or external table level?

    Hi,
    I have some data that I want to cleanse and transform. I don't want to cleanse it after i populate the external table, I want to get done with it at flat file level or while populating the external table. Can we cleanse and transform data at flat file or external table level through Oracle or OWB 11.2? Is it possible to run a conditional load (i.e. having a where clause or if-else-then) for an external table? Can we call oracle functions for an external table at the time of creation?
    Thanks in advance.
    Regards,
    Ann.

    Hi Oleg,
    Thanks a lot for the clarification. :)
    So is there a way that I can clease the data within the text file through Oracle or OWB? I have datatype mismatches in the data and most of my data is getting rejected because of that. The way I can think of, for solving this problem, is to create the external table with all fields with datatype varchar and then cleansing the data. But it doesn't seem very effecient plus it will get very complicated because I have almost 80-90 fields.
    Any help?
    Thanks and regards,
    Ann.

  • How to create a dynamic SSIS package for multiple flat file destinations

    Hi,
    I have to create a ssis package which has single data flow task and inside that I have 23 source (sql- select * from - statements)- destination (flat files, 23 distinct) connection.
    Now for each product I have to create separate SSIS package (i.e. if prod=abc then these read select * from abc_tables and 23 abc_ txt files)
    I want to do it dynamically, means only single package and inside that variables will take select * values for each source-dest connection (so i believe 23 variables) and same for destination flat files.
    Let me know. :)
    ANK HIT - if reply helps, please mark it as ANSWER or helpful post

    Sorry It seems you're contradicting yourself. you say I know my source and dest structure and the you're
    asking all I want is to have a dynamic structure
    what does that mean?
    and reading your next sentence
    I want to run a package for 5 products, instead of creating 5 ssis packages with 23 source- dest connection, I would
    like to have one with only 23 source- dest connections 
    What I feel is what you're looking for is to  have a looping structure to loop through each of the 5
    products.
    In that case what you could do is this
    1. Create a object variable in SSIS 
    2. Use a Execute SQL Task to populate the variable with all available products (I think you'll have a master table for that). Set ResultSet property to Full ResultSet and then in ResultSet tab map Object variable to 0 th index
    3. use a ForEachLoop container with ADO .NET recordset enumerator and map to object variable. Create a variable of datatype same as that of Product identifier field to get individual values out
    4. Inside loop create your data flow task with 23 source destination connection. In the query part use a parameter for product field and map it to the variable containing product value to get only data for the product.
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • Steps for performing Flat file to XML

    hey,
    does any one have steps for performing flat file (.csv) to XML conversion. how is the mapping in the design performed.
    kalyan.

    Kalyan,
    Consider my example, I have input file as csv structure and want it to convert into .xml file thats it.
    Input file
    J,24
    P,22
    I want the output file like
    <Emp_Details>
      <F1>J</F1>
      <F2>24</F2>
    </Emp_Details>
    <Emp_Details>
      <F1>P</F1>
      <F2>22</F2>
    </Emp_Details>
    I doesn't know whether the above matches exactly ur reqmt, but there is the option.
    Step 1 : Create Scenario & Business service
    http://www.flickr.com/photo_zoom.gne?id=699386732&size=o
    Step 2: Create sender & receiver comm.channel
    http://www.flickr.com/photo_zoom.gne?id=699386698&size=o
    http://www.flickr.com/photo_zoom.gne?id=699386664&size=o
    Step 3:Create all the objects in ID and cross verify below  whether u've created everything.
    http://www.flickr.com/photo_zoom.gne?id=699386690&size=o
    Step 4: Activate and run the interface.
    Your results :
    http://www.flickr.com/photo_zoom.gne?id=699386686&size=o
    Hope it helps!!!
    Best regards,
    raj.

  • SQL (Injection) Vs Flat File TXT Vs XML (XPath Injection) For Database.

    I will have to deal with robust login sessions so I need some advice on what type of database and language I will use.
    PHP and MySQL is the choice but I hear that flat file TXT are more faster so I am thinking about Perl.
    Perl is very flexiple and have many advantages but why do many folks use PHP.
    SQL and XPath injection are becoming more frequent plus MD5 and SH512 have been cracked by some angry folks.
    The Perl community is becoming very small so it will be difficult to receive support if I have a specific problem.

    In our projects we are extracting SMS messages and
    based on the
    MISDN (Telephone number) of each SMS we are fetching
    the corresponding
    account information (server name and password) from
    the database through (connecting through)
    webservice.The data is stored in the DB which you are accessing through a webserivce and that's too slow and unreliable so you are caching the results in a text file...
    That's horrible. The party responsible for the webservice should be fixing it or you should rethink the architecture. All this extra complexity and basically it's just causing problems for you. You'd be better off just accessing the DB directly. I assume this is possible since you know there is a database. In a good SOA architecture, you should not know or care about where the data is persisted.
    This webservice connection is slow/unreliable so a
    local copy of the
    MISDN number and corresponding account information
    are stored/updated in flat files tp prevent repeated
    accessing through webservice.
    My question is-which will be faster for accessing,
    using txt file or xml or a different format.I don't think it's going to matter much at all for the numbers you site, unless you are running this on a 20 year old PC or something.
    But I think you are addressing the wrong problem.

  • [SOLVED] Can't build fglrx module for 3.4.5-1-ARCH kernel

    Hello everyone!
    I updated my system using
        pacman -Syu --ignore glibc
    everything was good, I rebooted and my fglrx module isn't working!
    I tried:
        sudo catalyst_build_module
        Building fglrx module for 3.4.5-1-ARCH kernel ...
        Failed!!! Check out log: /var/log/catalyst-install.log
    Here is log
    Here is quote from it: Kernel header files are absent: directory /lib/modules/3.3.4-1-ARCH/build doesn't exist! Game over
    But I have the latest versions linux and linux-headers packages.
    Help please. Thanks.
    Last edited by metalex (2012-07-20 14:53:37)

    Yes, this is my problem, I  did all from here, but I still can't update glibc.
    grep '^lib/' /var/lib/pacman/local/*/files 
    output:
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/ld-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/ld-linux-x86-64.so.2
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libBrokenLocale-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libBrokenLocale.so.1
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libSegFault.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libanl-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libanl.so.1
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libc-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libc.so.6
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libcidn-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libcidn.so.1
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libcrypt-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libcrypt.so.1
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libdl-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libdl.so.2
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libm-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libm.so.6
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libmemusage.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnsl-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnsl.so.1
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnss_compat-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnss_compat.so.2
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnss_db-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnss_db.so.2
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnss_dns-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnss_dns.so.2
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnss_files-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnss_files.so.2
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnss_hesiod-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnss_hesiod.so.2
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnss_nis-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnss_nis.so.2
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnss_nisplus-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libnss_nisplus.so.2
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libpcprofile.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libpthread-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libpthread.so.0
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libresolv-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libresolv.so.2
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/librt-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/librt.so.1
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libthread_db-1.0.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libthread_db.so.1
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libutil-2.16.so
    /var/lib/pacman/local/glibc-2.16.0-1/files:lib/libutil.so.1
    /var/lib/pacman/local/lib32-glibc-2.15-10/files:lib/
    /var/lib/pacman/local/lib32-glibc-2.15-10/files:lib/ld-linux.so.2
    I don't know what to do with lib32-glibc...

  • Can I build an application for a linux machine from a windows machine?

    Is it possible, with LV2012, to build an application for a target other than the the machine type that I am working on?
    meaning, can I build an application to run on a linux machine from my development machine, which runs on windows?
    alternatively, can I build an applications to run (natively, no WOW) on a 32-bit OS from a development machine that is a 64-bit OS and 64-bit labview?
    Or does LV only build for similar targets as the development platform, and I would need to move the source code to another development machine to generate code for a different type of target machine?
    Thank you!
    Solved!
    Go to Solution.

    You can only do a build for the OS that you have a license for and for the OS bitness that you are using. In other words, you need LabVIEW for Linux or LabVIEW for Mac to build exe's for them and you need LabVIEW Windows 64-bit to build a 64 bit exe. It does require moving the source code to the other development systems.

  • How can I build Instalation Program for java applet?

    I want to build on line internet game based on java applet.
    Uset must instull the game to his computer.And this program mast Connect to the server when user enter
    to this application.Not from explorer,but from my application,like when i enter to ICQ.
    My questions:
    How can i build instalation program in java?
    How can i connect to the server when user enter into my game not from explorer but from game?
    Thanks,Maya.

    Hi Maya
    First of all an Applet does not have to be "installed" as you would with a normal application. The browser handles the install and execution of the applet.
    Check out this URL for more about applets:
    http://web2.java.sun.com/docs/books/tutorial/applet/index.html
    On the subject client/server pairs - see:
    http://web2.java.sun.com/docs/books/tutorial/networking/sockets/index.html
    If you decide to make the game an application instead check out Sun's WebStart - under the products section.
    That's about all the help I can give you.
    Programming a game is a complicated subject - I am doing one too:
    http://hjem.get2net.dk/mhjembaek/armageddon/
    Kind reg,
    Hjembaek

  • Can I use multiple values for a rollup key on the same Endeca record?

    We have a business need to to aggregate our records using different criteria, based on user navigation. We are thinking of using a rollup key with multiple values to help with the aggregation, but we are running into some issues.
    Here is a made-up xample of what we want to do: assume we have a group of products and these products can be organized into groups using parent-child relationships. We would like to create an aggregate record for each parent, and we want the aggregate record for each parent to include all the children for the parent. To achieve this, we use a field called "parent_rec_spec" that holds the parent record spec and we set the same value on the field for the parent and its children. When we do rollup using the parent_rec_spec as the rollup key, we are able to see one aggregate record for each parent (with its children).
    The previous setup worked perfectly for us so far. But now we are getting a business requirement that allows children nodes to be linked to multiple parents at the same time. We were hoping of using another dimension to limit the records based on user roles/characteristics , so that only applicable parents/children are displayed (for example, we can use "market" as an additional filtering property, and we decide to show all parents/children for "North America", while hiding the parents/children for other markets).
    This caused an odd behavior when children are linked to multiple parents. For example, assume that SKUs A and B were linked to parents in "North America" and "Europe" at the same time, and assume that the user chose the "North America" market. The navigation state would eliminate the parents/children that are no in North America, and will also cause the parents/children that are labeled for North America to show up and be aggregated correctly. This however will lead to the creation of additional aggregate records for the A and B using the parent_rec_spec values that would have linked them to the Europe parents (even though the parents are hidden away).
    Here is an example index file that we used to load the test data:
    Update||1
    Market||North America
    Record Type||Product
    Name||Parent 1
    rec_spec||P1
    parent_rec_spec||P1
    EOR
    Update||1
    Market||Europe
    Record Type||Product
    Name||Parent 2
    rec_spec||P2
    parent_rec_spec||P2
    EOR
    Update||1
    Market||North America
    Record Type||Product
    Name||Child A
    rec_spec||A
    parent_rec_spec||P1
    EOR
    Update||1
    Market||North America
    Market||Europe
    Record Type||Product
    Name||Child B
    rec_spec||B
    parent_rec_spec||P1
    parent_rec_spec||P2
    EOR
    Update||1
    Market||North America
    Market||Europe
    Record Type||Product
    Name||Child C
    rec_spec||C
    parent_rec_spec||P1
    parent_rec_spec||P2
    EOR
    Update||1
    Market||Europe
    Record Type||Product
    Name||Child D
    rec_spec||D
    parent_rec_spec||P2
    EOR
    In this setup, we have parent P1 marked for North America with children A, B and C, and parent P2 marked for Europe with B, C and D as children. When we use North America as a filter in the navigation state, and parent_rec_spec as the rollup key, then we will see an aggregate record for P1, A, B and C. But we will also see an aggregate record for B and C by itself (presumably because of the other parent_rec_spec value on these records).
    The actual data that we are testing with is more complicated, but the end result is similar. We also noticed that the additional aggregate records would not be created always, depending on the ordering of the records.
    The question that I need help with is this: is there a way to fine tune the rollup logic so that it can only include certain records (in the example above, we can change the rec_spec from PA and PB to PA_North_America and PB_Europe and then we would be interested in rolling up using values that end with NorthAmerica).
    By the way, we considered using separate rollup keys for each context (like parent_rec_spec_north_america and parent_rec_spec_europe), but the number of contexts is dynamic, and might grow large. So it is not easy for us to create the additional properties on the fly, and we are concerned about the possible large number of dimensions.

    http://www.adobe.com/cfusion/webforums/forum/messageview.cfm?catid=2&threadid=1157850

  • Can we give UNIQUE ACCESS FOR THE SPECIFIC FILE IN THE LIBRARY in SP2013? How can we remove users from SHARED WITH link where files are shared with users?

    Hi,
    Any help on this?
    Thanks
    srabon

    Hi srabon,
    For giving unique access for a specific file in a library, you can go to the library, and select the file , and click FILES->Shared With->ADVANCED, under PERMISSION ribbon, click ‘Stop Inheriting Permissions’, then the file will have unique permissions.
    For removing the shared users for a file, firstly, like the above steps, select the file , and click FILES->Shared With->ADVANCED , make sure the file has unique access, then select the users that you want to remove, and click Remove User Persmissions
    under PERMISSIONS ribbon.
    I hope this helps.
    Thanks,
    Wendy
    Wendy Li
    TechNet Community Support

  • How can i set a path for my deployment files in weblogic server 10.3

    Hi
    How can i set the path for my WAR ,JAR files while deploying.i am using the wls10.3 version.
    is there any scripts for this ,please provide me.
    my Application is ADF 11g application.

    By "path", I assume you mean "classpath".
    The simplest way is simply to include the jars you need inside the web application or web module's WEB-INF/lib directory, EJB module's META-INF/lib directory, or EAR lib directory.
    If that's not practical, if you use NodeManager to start your servers, you can go to the "Server Start" tab in the server definition in the WebLogic console and edit the "Classpath" field, which defaults to no value. You can specify a classpath value there. Note that if you specify a value there, it REPLACES the default classpath for the server, it doesn't add to it. If you need to just add to it (a much more likely scenario), if the value references the value "$CLASSPATH" in it, that will reference the original classpath value that the server would have had.
    So, for instance, if you wanted to include the MQ jars in the server classpath, you could set a value like this:
    /usr/java/mq/lib/mq.jar:/usr/java/mq/lib/mqstuff.jar:$CLASSPATH

  • Can you block your number for texting, so dont show in ppls phone record?

    usually in the phone record show the number who is text you and the phone you sent a message. But since a week a go don't show the number of the person, and I received tons of messages everyday, but  on my phone record of text messages don't reflect the phone number. Is a way you can block your number??? so dont registrer in the phone record?

        I understand the importance of viewing numbers in your message details hmaldonado. There is an option for individuals to block their number if it is restricted. What do you see on your phone for these messages you are referring to? Is this a number that is stored in your phonebook? Have you checked My Verizon online? Can you see calls from this person? What specific task are you trying to accomplish with this number? Here's how to access My Verizon. Select My Usage, then Usage details. http://bit.ly/xB4iTc
    JonathanK_VZW
    VZW Support
    Follow Us on Twitter@VZWSupport

Maybe you are looking for

  • No Internet Access ASA 5510 with Site to Site VPN

    I have inherited an ASA5510. We have configured a site to site VPN between it and a Juniper SRX550. The Site to Site VPN works fine but users behind the ASA can not see the Internet. I'm pretty sure I'm missing something. Any help would be greatly ap

  • Running rwservlet with in-process server in JDeveloper

    Hello, I have a report.jsp in JDeveloper that contains both a web and paper based layout. I want to run this report within my web application using the in-process server. From what I understand, if I want PDF format I have to use the rwservlet. When

  • Converting excel document into a tempravory database

    i have to acept a excel document and convert it to the tempravory database and allow the user to edit .if the user have changed anything it should change the details in the excel document. anybody know how to do this..

  • Do I need additional Driver/Clients when accessing Oracle from CSharp ?

    I am trying to connect from within a CSharp program to Oracle database 10g Express. I coded in CSharp: using System.Data.OracleClient; but this line gives an unresolve error: Type or namespace missing ...? As far as I thought VisualStudio 2005 provid

  • Use of 0CRMVERSION in the coupon planning

    Hi Friends, What is the significane of the Infoobect 0CRMVERSION in the CRM coupon planning do we have any specific values for this 0CRMVERSION. Please share your ideas on this. This is used in our system BPS coupon planning for CRM integration proje