digital communication basic

Upload: pankaj-jadhav

Post on 14-Apr-2018

221 views

Category:

Documents


0 download

TRANSCRIPT

  • 7/30/2019 Digital Communication Basic

    1/12

    9/3/13 Digital Communication

    members.chello.nl/~m.heijligers/DAChtml/digcom/digcom.html 1/12

    Digital Communication

    Introduction

    In Figure 1 an overview of the typical functional elements of a digital communication system are shown.

    Figure 1: Digital Communication System

    The information source generates particular symbols at a particular rate. The source encoder translates these

    symbols in sequences of 0's and 1's. The channel encoder is oriented towards translating sequences of 0's

    and 1's to other sequences of 0's and 1's, to realize high transmission reliability and efficiency. The modulator

    accepts streams of 0's and 1's, and converts them to electrical waveforms suitable for transmission.

    The communication channel provides the electrical connection between the source and destination. It has a

    finite bandwidth, and the waveform transmitted suffers from amplitude distortion and phase distortion. In

    addition to distortion, power is decreased due to attenuation of the channel. Finally, the waveform is

    corrupted by unwanted electrical signals, referred to as noise. The primary objective of a communication

    system is to suppress the bad effects of noise as much as possible.

    The inverse process takes place at the destination side. The demodulator converts the electrical waveforms to

    sequences of 0's and 1's, the channel decoder translates the sequence of 0's and 1's to the original sequence

    of 0's and 1's. It also performs error correction and clock recovery. The source decoder finally translates the

    sequence of 0's and 1's into symbols.

    The gray boxes in the following figure show how the elements are grouped in a digital audio configuration, in

    this case a CD player and a DA-convertor.

    Information sourceThe main digital source for consumer purpose is the CD-player. This section will present a short summary

    about the history and principles of a CD-player. It is derived from a document of Grant M. Erickson, e.g. see

    A Fundamental Introduction to the Compact Disc Player.

    http://www.tc.umn.edu/~erick205/Papers/paper.html
  • 7/30/2019 Digital Communication Basic

    2/12

    9/3/13 Digital Communication

    members.chello.nl/~m.heijligers/DAChtml/digcom/digcom.html 2/12

    The history of the CD player

    As staggering as the release of the compact disc player was in 1982, the technology and theories which

    allowed it to be born were long in development. In 1841, the great mathematician Augustin-Louis Cauchy

    first proposes the sampling theorem. Nearly 80 years later J.R. Carson publishes a mathematical analysis of

    time sampling in communications. In a 1928 lecture at the American Institute of Electrical Engineers Harry

    Nyquist provides proof of the sampling theorem in "Certain Topics in Telegraph Transmission Theory". In

    1937, A. Reeves proposes pulse code wave modulation (PCM). In 1948, John Bordeen, William Shockley,and Walter Brattain invent the bipolar junction transistor at Bell Labs--compact digital circuitry is a reality.

    Two years later, in 1950 Richard W. Hamming publishes significant work on error correction and detection

    codes. In 1958 C.H. Townes and A.L. Shawlow invent the laser. In 1960 R.C. Bose publishes binary group

    error correction codes. That same year I.S. Reed and G. Solomon publish error correction codes to be used

    in the CD player 22 years later. Also early computer music experiments take place at Bell Labs. Fifteen years

    before consumers see the first player, NHK Technical Research Institute publicly demonstrates a PCM digital

    audio recorder with a 30 kHz sampling rate and 12-bit resolution. Two years later, Sony Corporation

    demonstrates a PCM digital audio recorder with a 47.25 kHz sampling rate and 13-bit resolution. A

    hemisphere away, Dutch physicist Klaas Compaan uses a glass disc to store black and white holographicimages using frequency modulation at Philips Laboratories. Four years later, in 1973 Philips engineers begin

    to contemplate an audio application for their "video" disc system. A prototype disc with a 44 kHz sampling

    rate is run through a 14-bit digital-to-analog converter and exhibits a signal-to-noise (S/N) ratio of 80 dB in

    monaural. Now a research frontier, Mitsubishi, Sony, and Hitachi all demonstrate digital audio discs at the

    Tokyo Audio Fair in 1977. One year later, Philips joins with its recording subsidiary Polygram Records to

    develop a worldwide digital audio standard. In March 1979, Philips demonstrates a prototype compact disc

    player in Europe. Sony joins the Philips/Polygram coalition after Matsushita declines. In June of 1980, the

    coalition formally proposes their CD standard. A year later in 1981, Sharp successfully mass produces the

    semiconductor laser. This step was crucial to delivering a consumer product. In Fall of 1982 nearly 150 years

    of work comes to fruition and Sony and Philips introduce their respective players to consumer in Europe. The

    following spring, the player is introduced in the United States.

    The specifications for the compact disc and compact disc players were jointly developed by Sony, Philips,

    and Polygram as mentioned previously. This specification is contained in their standards document referred to

    as the Red Book. Some of these specifications are summarized in Table 1, 2 and 3.

    Table 1: Disc format

    Playing time 74 minutes, 33 seconds maximum

    Rotation 1.2 - 1.4 m/sec. (constant linear velocity)

    Track pitch 1.6 m

    Diameter 120 mm

    Thickness 1.2 mm

  • 7/30/2019 Digital Communication Basic

    3/12

    9/3/13 Digital Communication

    members.chello.nl/~m.heijligers/DAChtml/digcom/digcom.html 3/12

    Center hole diameter 15 mm

    Recording area 46 mm - 117 mm

    Signal area 50 mm - 116 mm

    Material Any acceptable medium with a refraction index of 1.55

    Minimum pit length 0.833 m (1.2 m/sec) to 0.972 m (1.4 m/sec)

    Maximum pit length 3.05 m (1.2 m/sec) to 3.56 m (1.4 m/sec)

    Pit depth ~0.11 m

    Pit width ~0.5 m

    Table 2: Optical system

    Standard wavelength 780 nm

    Focal depth 2 m

    Table 3: Signal format

    Number of channels 2 channels (4 channel recording possible)

    Quantization 16-bit linear

  • 7/30/2019 Digital Communication Basic

    4/12

    9/3/13 Digital Communication

    members.chello.nl/~m.heijligers/DAChtml/digcom/digcom.html 4/12

    Quantizing timing Concurrent for all channels

    Sampling frequency 44.1 kHz

    Channel bit rate 4.3218 Mb/sec

    Data bit rate 2.0338 Mb/sec

    Error correction code Cross Interleave Reed-Solomon Code (with 25% redundancy)

    Modulation system Eight-to-fourteen Modulation (EFM)

    The compact disc player contains two main subsystems: the audio data processing system and the

    servo/control system. The servo, control, and display system orchestrate the mechanical operation of the

    player and include such items as the spindle motor, auto-tracking, lens focus, and the user interface. The

    audio data processing section covers all other player processes.

    Source encoding/decoding

    Digital data provided by digital sources like CD, DAB, DAT or DCC is transmitted by using the SPDIF

    consumer standard, derived from the AES/EBU professional standard. Using this standard, every sample is

    transmitted as a 32-bit word (called a subframe). Two subframes make one frame (64 bits total), which

    repeats at the sampling rate in use. A block contains 192 frames. At 48 kHz the bit rate will be 3.072 MHz,

    at 44.1 kHz the bit rate will be 2.8224 MHz. The first subframe will contain the sample from channel A, the

    second frame will contain the sample from channel B.

    The basic SPDIF and AES/EBU subframe structure can be found in the following figure:

    Figure 2: SPDIF subframe structure

    The first 4 bits in the preamble are used for synchronization. There are 3 different sync-patterns (called B, M,

    W), but they can appear in different forms, depending on the value of the last cell of the previous subframe

    (see the section about channel encoding/decoding for more information). Preamble B marks a wordcontaining data for channel A (left) at the start of a block. Preamble M marks a word with data for channel A

    that isn't at the start of a block. Preamble W marks a word containing data for channel B (right, for stereo).

  • 7/30/2019 Digital Communication Basic

    5/12

    9/3/13 Digital Communication

    members.chello.nl/~m.heijligers/DAChtml/digcom/digcom.html 5/12

    Figure 3: Block structure

    There is room for 24 audio bits (4-27) in a subframe, but normally 20 audio bits are available (8-27). In case

    of CD only 16 bits are used. Bits that are not used, are defined to be zero in SPDIF (in contrary to for

    instance I2S, where the LSB is repeated, resulting in less transitions in a serial interface, and hence less power

    dissipation). As audio data must be in 2-complements code, and different word lengths may be used, the

    MSB must always be at the same place. SPDIF sends the MSB first.

    Four status bits (28-31) accompany each subframe. Bit 28 (V) is the validity flag, which indicates whether the

    data received is suitable for conversion to an analog signal (e.g. to indicate non-audio data, like CD-I or CD-

    ROM players, which can be used for muting the audio outputs). Bit 31, the parity bit (P), produces even

    parity (total number of ones is even), which is meant for error correction at the destination. The user bit (bit

    29, U) and channel bits (bit 20, C) are collected to form user and channel status blocks.

    The sequence of channel bits over 192 subframes (or one block) builds up a 24 byte (192 bit) channel status

    block. The preamble is used to denote the start of a channel status block (see the section about channel

    encoding/decoding for more information). The structure of a channel status block is shown in the following

    figure.

    Figure 4: Channel status block structure

    First, the serial data is assembled into 12 words of 16 bits each. The first 6 bits of the first word form a

    control word. The following 2 bits permit a mode select for future expansion. Currently only mode 0 is

    standardized (and the corresponding bit allocation shown in the figure).

  • 7/30/2019 Digital Communication Basic

    6/12

    9/3/13 Digital Communication

    members.chello.nl/~m.heijligers/DAChtml/digcom/digcom.html 6/12

    Control word: bit 0 0 = consumer, 1 = professional

    bit 1 0 = normal, 1 = digital data

    bit 2 0 = copy prohibit, 1 = copy permit

    bit 3 0 = no pre-emphasis, 1 = pre-emphasis

    bit 4 reserved

    bit 5 0 = 2CH, 1 = 4CH

    Mode: bit 6,7 00 = mode 0 (01 10 11 = reserved)

    Category code: bit 8, 15 00000000 = general format

    10000000 = CD player

    11000000 = DAT player

    bit 15 0 = original recording

    1 = first-generation copy (SCMS)

    Source no.: bit 16 - 19 0000 = don't care

    0001 = source 1

    0010 = source 2

    ...

    Channel no.: bit 20 - 23 0000 = don't care

    1000 = A (left channel for stereo)

    0100 = B (right channel for stereo)

    1100 = C

    ...

    Sampling rate: bit 24 - 27 0000 = 44.1 kHz

  • 7/30/2019 Digital Communication Basic

    7/12

    9/3/13 Digital Communication

    members.chello.nl/~m.heijligers/DAChtml/digcom/digcom.html 7/12

    0100 = 48 kHz

    1100 = 32 kHz

    remaining codes reserved

    Clock accuracy: bit 28, 29 00 = normal accuracy

    10 = high accuracy

    01 = variable speed

    In case of a CD player (category code 10000000), the audio sample words in the subframe structure

    consists of 16 bit words, MSB in position 8, bits 24 - 27 in the channel status block are 0000, indicating a

    44.1 kHz sampling frequency, and bits 28,29 are set according to the source accuracy. The two channel

    format (00000000) leaves room for a maximum of 20 bits/sample, and various sample frequencies.

    W.r.t. possible sample rates and pre-emphasis the following observations can be made. CD uses 44.1kHz,

    with the possibility of 50/15 s pre-emphasis. DAB uses 48 kHz, without pre-emphasis. DCC and DAT

    support 32, 44.1, 48 kHz with the possibility for 50/15 s pre-emphasis. ISO/MPEG supports 6 sampling

    rates (16, 22.05, 24, 32, 44.1, 48 kHz) with 2 possible pre-emphasis-types (not simultaneously): 50/15 s

    and the so-called CCITT J.17. CD-i only uses 44.1 kHz (also for Full-Motion Video with MPEG), with the

    possibility for 50/15 s pre-emphasis. The whole 50/15 s pre-emphasis matter is a gift from Sony, who in

    the early digital period weren't capable of building reasonable DA-convertors. The CCIT standard originates

    from the telecommunication environment.

    The user code bits can be used by the manufacturer at will. They are used in blocks of 1176 bits before

    which a sync-word of 16 "0"-bits is transmitted.

    Channel encoding/decoding

    Transmitting a digital signal by simply serializing the data is not a good strategy, since it is difficult to estimate

    the data rate of the source at the destination, and hence it is difficult to separate the received bits and to

    recover an accurate jitter-free clock.

    PDIF interfaces use FM channel coding (also known as Manchester coding or bi-phase mark coding), which

    is DC free, strongly self-clocking and capable of working with a changing sampling rate. The use of FM

    means that the channel frequency is the same as the bit rate when sending data ones, because FM takes care

    that a state transition takes place at each bit. A logical 1 is represented by a second transition. Hence the

    clock signal can be recovered, independent of the data.

    Table 4: Biphase mark coding examples

    data Biphase mark equivalent

    0000 00110011

  • 7/30/2019 Digital Communication Basic

    8/12

    9/3/13 Digital Communication

    members.chello.nl/~m.heijligers/DAChtml/digcom/digcom.html 8/12

    0111 00101010

    0101 00101101

    The sync-pattern deliberately violates the Manchester coding, so that the beginning of a block is easy todetect. The bit patterns in the preamble always contain more than two 0's or 1's in a row.

    Table 5: Preamble contents

    Sync pattern

    Contents

    (bit 31="0")

    Contents

    (bit 31="1")

    B 11101000 00010111

    M 11100010 00011101

    W 11100100 00011011

    Modulation/demodulation

    The electrical interface of the consumer format as specified by IEC 958 is shown in the following figure.

    Figure 5: Electrical SPDIF interface

    It uses a 0.5 Vtt signal, and a frequency of almost 6 MHz1, which can be conveyed down conventional

    audio-grade coaxial cable connected with RCA phono plugs.

    A slicer can be used to convert the analog waveform to a binary output. The signal voltage is compared to a

    particular voltage, called the threshold, and if the signal voltage is above the threshold the comparator outputsa high level, and a low level otherwise. If the waveform would contain a DC component, the average level of

    the waveform would raise, and hence this would seriously affect the slicing process. Hence it is clearly not

    possible to serialize arbitrary data in a shift-register for direct transmission. The manchester coding process

    http://members.chello.nl/~m.heijligers/DAChtml/digcom/digcom.html#pgfId-998574
  • 7/30/2019 Digital Communication Basic

    9/12

    9/3/13 Digital Communication

    members.chello.nl/~m.heijligers/DAChtml/digcom/digcom.html 9/12

    takes care that on average the number of 0's is equal to the number of 1's, making slicing possible.

    Communication channel

    The communication channel consists of a coax-cable, connecting the source and destination. The waveform

    transmitted through the cable is affected by noise, baseline wander, intersymbol inerference and imperfect

    equalization, resulting in time uncertainties in the position of the edges, resulting in jitter when a slicer is used to

    recover the clock. An eye-pattern can be used to observe the characteristics of the communication channel.

    In Figure 6 it is shown how particular waveforms (left hand side) are typically displayed using an oscilloscope

    (right hand side) triggering on such a waveform.

    Figure 6: Eye diagrams (unlimited bandwidth ideal, limited bandwidth ideal, distorted)

    In Figure 7 the typical characteristics derived from such an eye-pattern are shown. Noise closes the eye in a

    vertical direction, whereas jitter closes the eye in a horizontal direction.

    Figure 7: Eye diagram characteristics

    In [Watk94] the following minimum eye pattern (see [Shan85]) acceptable for correct decoding of data for a

  • 7/30/2019 Digital Communication Basic

    10/12

    9/3/13 Digital Communication

    members.chello.nl/~m.heijligers/DAChtml/digcom/digcom.html 10/12

    biphase-marked data stream in the AES/EBU standard is given (see Figure 8). In this figure Tnom denotes

    the half of a biphase symbol period, and Tmin = 0.5 * Tnom . Because in the biphase-marked digital data

    stream a data bit is represented by two transitions, the jitter margin can only be a half bit (in the absence of

    noise) to be able to reconstruct the data, which is about 80ns. The minimum height of the eye should be

    200mV.

    As error correcting methods are incorporated inside the SPDIF interface, data errors because of data

    corruption can be detected very easily. In normal conditions, data corruption caused by jitter or noise ishardly an issue.

    Figure 8: SPDIF eye pattern

    Cable impedance

    The coax-cable has a so called impedance,which will be explained in the remainder of this section.

    Transmission can be modelled as electromagnetic energy travelling from one place to another. Depending on

    the frequency of the of the electromagnetic energy, different characteristics can be observed when it is

    tranported by conductors.

    An electric model of a conductor can be found in Figure 9.

    Figure 9: Electrical model of coax cable

    In a nutshell the cable capacitance Cresults from the fact that there exists an electric field between the two

    conductors given a potential difference between them, determined by the geometry of the space between the

    conductors and the nature of the dielectric. The cable inductance L results from the fact that any current-

    carrying conductor produces a magnetic field. A cable (with the exception of superconductors) always has a

    certain amount of resistance R . And finally, the insultor will cause signal losses modelled by G .

    For an infinite length of the network model we find:

  • 7/30/2019 Digital Communication Basic

    11/12

    9/3/13 Digital Communication

    members.chello.nl/~m.heijligers/DAChtml/digcom/digcom.html 11/12

    The equation does not depend on the length, because these quantities are distributed. No matter how short

    the length of the cable is, it still can be modelled as above. The cable impedance is only a concern when the

    cable is a fraction of the wavelength of the signal in question.

    At low frequencies electromagnetic energy is called electricity. Electricity is transported completely inside

    conductors. At DC an inductor acts as a short circuit, and a capacitance acts as an open circuit. The

    resistance is determined by the cross-sectional area of the conductor. The insulation has no effect on the

    ability of the conductor to pass current.

    As frequency rises, inductors display an increasing impedance with frequency and capacitors show falling

    impedance. Electromagnetic energy becomes increasingly desperate to leave the conductor. This results in the

    skin-effect, where current only flows in outside layer of the inductor, causing the resistance to rise. As the

    energy is starting to leave the conductor, the geometry of the space between them is becoming more

    important. This determines the impedance. A change of impedance causes reflections in the energy flow,

    hence constant impedance cables with fixed inductor spacing are necessary. As frequency rises even further,

    the energy travels less in the conductors and more in the insulation between them. The composition in the

    insulators is becoming more important, and they begin to be called dielectrics. Poor dielectrics like PVC

    absorb high-frequency energy, and attenuate the signal.

    As the transition rate is within an average of a few megabytes per second, and only moderate cable lengths

    will be used in practice (hence R and G can be ignored to a certain extent), the main losses of the cable will

    be determined byL and C :

    Notice that the frequency dependency is gone. Also notice that the result does not depend on the cable length

    (when ignoring cable losses), because the quantities are distributed along the cable (you can't isolate the C-

    part or L-part from a cable). Also notice that the bandwidth of the signal should be narrow, and the transition

    of a sequence of many identical bits should therefore be avoided by modifying the spectrum of the signal by

    application of a suitable channel encoding.

    So, if the cable length is a fair fraction of the wavelength of the signal in question, the cable can be considered

    as a transmission line, and pulses travel down it as current loops. If a positive pulse is launched across the

    line, it will charge the dielectric in the direction from the driver to the receiver, resulting in a leading edge.

    When the driver ends the pulse, the dielectric will discharge leading to a falling edge, which bears the same

    current as the leading edge. There is thus a loop of current rolling along the line as if it were a Caterpillartractor.

    Transmission lines which transport energy in this way have a characteristic impedance because of the interplay

    of inductance along the conductors with the capacitance. One consequence is that a correct termination or

    matching is required between the line and both the driver and the receiver to make the rolling energy roll

    straight out of the line into the load. If the impedance is mismatched, reflections will result.

    The main concern with digital cables is of the junction that forms at the major connection. If the connectors

    are not clean or the cables are not connected in a right manner, a diode effect can be observed, resulting in a

    slow rise time and a fast fall time.

    1. More exactly, one frame contains 64 bits which are sent during one clock period. This results in a bit-rate

  • 7/30/2019 Digital Communication Basic

    12/12

    9/3/13 Digital Communication

    of 2.8224Mbit/s (44.1kHz), 3.072 Mbit/s (48kHz), or 2.048Mbit/s (32kHz), which after bi-phase mark

    coding is doubled.

    Copyright 2001, Marc Heijligers and the DAC group - All rights reserved.